Course KI-Campus-Originale

Transformer-basierte Large-Language-Models verstehen

This course provides a conceptual understanding of modern large language models. It focuses on the Transformer architecture and its key components, such as attention and positional encoding. With this knowledge, your will gain insight into the inner workings of current models from OpenAI and DeepSeek and understand how they function.

Transformer basierte LLMs
📊︎ Intermediate
10 Stunden
🏅︎ Record of Achievement
🎁︎ For free
© CC BY-SA 4.0
🌐︎ German

Overview

Large language models are now central to many AI applications, from integrated chatbots to AI agents. But how do these systems actually work?

In this online course, you’ll receive a clear introduction to the fundamentals of modern language models. The focus is on the Transformer architecture, which forms the basis of today’s large language models. Your will become familiar with key concepts such as attention, positional encoding, and the modular structure of the architecture.

The course emphasizes a clear conceptual understanding and deliberately avoids programming and complex mathematical derivations. The most important ideas are explained step by step using clear examples.

Building on this foundation, we will then examine current models from OpenAI and DeepSeek, looking at how they are structured, how they have evolved in recent years, and how they differ from one another.

Which topics will be covered?

  • Fundamentals of modern large language models and their role in current AI applications
  • Structure and functioning of the Transformer architecture
  • Structure and development of modern model families such as GPT and DeepSeek

What will I achieve?

By the end of the course, learners will be able to:

  • explain how large language models operate and what they are utilized for.
  • describe the key components of the Transformer architecture.
  • understand the development and structure of current model families such as GPT or DeepSeek.
  • compare different LLM architectures based on fundamental criteria. 
     

Which prerequisites do I need to fulfill?

A basic understanding of neural networks or an introductory course on topics such as tokenization, embeddings, or chatbots is helpful but not required

 

The underlying project was funded by the Federal Ministry of Research, Technology and Space under the funding code “KI-Servicezentrum Berlin-Brandenburg” 16IS22092. Responsibility for the content of this page remains with the author.

This course is offered by

KI-Servicezentrum
Gefördert durch: BMFTR

Helpdesk

Hint: Have you already checked our FAQ for an answer to your question?