Lesson 10 of 2116
The Three Ingredients: Data, Compute, Algorithms (Capstone)
Every AI breakthrough of the past decade rests on three interacting ingredients. Synthesize everything you have learned into one working model.
Lesson map
What this lesson covers
Learning path
The main moves in order
- 1The Ingredient Model
- 2data
- 3compute
- 4algorithms
Concept cluster
Terms to connect while reading
Section 1
The Ingredient Model
At the broadest possible level, every AI system in history is a mix of three ingredients. Data provides the examples. Compute runs the calculations. Algorithms decide how the calculations relate to each other. Shifting any one of them creates a new era.
The ingredients mapped to history
Compare the options
| Era | Data | Compute | Algorithm |
|---|---|---|---|
| 1980s expert systems | Handwritten rules | Desktop CPUs | Rule-based inference |
| 2000s statistical ML | Small curated datasets | Commodity servers | SVM, decision trees |
| 2010s deep learning | ImageNet-scale labeled sets | GPUs in parallel | Convolutional and recurrent nets |
| 2020s LLM era | Internet-scale text + human feedback | Datacenter-scale clusters | Transformer with RLHF |
| 2025-2026 reasoning era | Synthetic + verifiable traces | Test-time inference scaling | RL on self-generated reasoning |
How the ingredients interact
- Data bottlenecks: public text ran out, driving synthetic and multimodal data
- Compute bottlenecks: GPU supply and energy availability cap scaling
- Algorithm bottlenecks: pure transformers plateau, pushing MoE, state space, and hybrid architectures
- Adding one ingredient often just exposes a shortage of another
Walkthrough: train your own tiny LLM mentally
- 1Choose your data: 1B tokens of curated web + code
- 2Choose your compute: 8 H100 GPUs for 72 hours
- 3Choose your architecture: 1B-parameter decoder transformer with rotary embeddings and grouped-query attention
- 4Tokenize the corpus with a 32k-token BPE vocabulary
- 5Train with AdamW, cosine learning-rate schedule, 2048-token sequences
- 6Evaluate on held-out perplexity and a small MMLU slice
- 7Fine-tune on 50k human-rated instruction pairs
- 8Apply RLHF or DPO for a final style pass
- 9Deploy via vLLM with structured output support
- 10Monitor for drift and retrain quarterly
Every decision here is a bet on one of the three ingredients.
# Minimal training config sketch
config = {
"model": {
"layers": 24,
"d_model": 2048,
"n_heads": 16,
"kv_heads": 4,
"vocab_size": 32000,
"max_seq_len": 2048,
"position_embedding": "rope",
},
"train": {
"tokens": 20_000_000_000, # ~20 tokens per param, Chinchilla-ish
"batch_size": 512,
"optimizer": "adamw",
"lr": 3e-4,
"schedule": "cosine",
},
"data": {
"sources": ["fineweb", "the_stack_v2", "textbooks_synthetic"],
"dedup": True,
"language_filter": "en",
},
}Capstone exercise
- Pick a real AI product you use (chat, code assistant, image generator).
- Estimate which ingredient is its main competitive advantage and why.
- Predict what will change first for that product: the data, the compute, or the algorithm.
- Identify what signal in the wild would confirm your prediction within 12 months.
- Write it down and revisit it later. You will be grading yourself on real-world feedback.
“Everything in AI is ultimately a budget problem with three line items.”
Key terms in this lesson
The big idea: AI is not magic. It is three ingredients, thousands of engineers, and careful attention to which ingredient is scarce this year. You now have the vocabulary, the history, and the mental model to reason about whatever comes next.
End-of-lesson quiz
Check what stuck
15 questions · Score saves to your progress.
Tutor
Curious about “The Three Ingredients: Data, Compute, Algorithms (Capstone)”?
Ask anything about this lesson. I’ll answer using just what you’re reading — short, friendly, grounded.
Progress saved locally in this browser. Sign in to sync across devices.
Related lessons
Keep going
Creators · 45 min
Open vs. Closed Models: Philosophy and Strategy
Open-source AI is both a technical movement and a political one. Understand the arguments so you can pick a stack and defend it.
Builders · 35 min
A Short History: From Expert Systems to Transformers
AI did not start in 2022. It has decades of wrong turns and breakthroughs. Knowing the history helps you spot hype from real progress.
Creators · 50 min
The Full Machine Learning Pipeline
From raw bytes to deployed model, every ML system follows the same ten-stage pipeline. Master it and you can read any architecture paper.
