For educators

Hero image: For educators

Lesson plan 1: LLMs Unplugged Fundamentals

This core workshop covers the essential training-to-generation pipeline. Start with a brief introduction to set the scene, then move through training a model, generating text from it, and finally exploring what happens when you use a larger pre-trained model. Each step builds on the last, giving students a complete picture of how language models work.

Suggested timing

Notes for Fundamentals

Lesson plan 2: going deeper

For students ready to go further, this extended trajectory adds the “how models understand” topic. After covering the fundamentals, you explore how models can track grammatical context and how words get represented as numerical vectors. This path suits later-year high school students, computing electives, or keen beans who want to understand what “ Attention mechanism The ability to focus on relevant previous words when making predictions. In real LLMs, attention is learned, weighted, and dynamic---the model decides what to focus on for each prediction. Context columns illustrate the motivation for attention: considering more than just the immediately preceding word. View in glossary ” and “ Embedding A numerical representation of a word. Each row in your bigram grid is that word's embedding vector---a fingerprint of its usage context. In real LLMs, embeddings are learned separately rather than derived from raw counts, but the principle is the same: words used in similar ways get similar vectors. View in glossary ” actually mean.

What these additions cover

Why split the trajectory?

The fundamentals work for any audience and require only 90 minutes. The “understanding” lessons require more time and comfort with abstraction, but they connect directly to concepts students will encounter in any deeper study of AI: attention mechanisms, embeddings, vector similarity. Running them as a second session (or a follow-up for interested students) keeps the core workshop accessible while offering a clear path forward.

Lesson plan 3: controlling output

Once students can generate text, a natural question is: “How do you make it more or less creative?” The sampling lesson shows how temperature and truncation strategies change the character of output without changing the model itself. This is a quick add-on to either the fundamentals or the deeper trajectory.

This lesson explains:

Students discover that “creativity” in AI comes from two controls: adjusting probability distributions and filtering which tokens to consider. The same model can produce cautious prose or wild poetry just by tweaking these parameters.

Adaptation and data

For classes focused on data science, ethics, or media literacy, the “Adaptation and data” topic explores what happens when models train on their own output.

The Synthetic data lesson is particularly effective for discussions about:

This works well as a standalone activity after students have done basic training and generation, or as part of a broader unit on AI ethics and media literacy.