Lesson 227 of 1570
Reasoning Models: OpenAI o1 and After
In 2024, a new class of models traded fast answers for slow, deliberate thinking, and benchmarks jumped.
Lesson map
What this lesson covers
Learning path
The main moves in order
- 1Thinking Longer, On Purpose
- 2reasoning models
- 3o1
- 4inference-time compute
Concept cluster
Terms to connect while reading
Section 1
Thinking Longer, On Purpose
In September 2024, OpenAI previewed o1, a model that spent extra compute before answering, generating long internal chains of reasoning. On hard math, coding, and science benchmarks, o1 leapt past GPT-4o, sometimes by double-digit points on tests where progress had been crawling.
The core idea was not prompt-level chain of thought. It was training the model, often through reinforcement learning, to use its own thinking tokens effectively, and then letting it spend as many of those tokens as needed at inference time.
What reasoning models do well
- Multi-step math and proofs where intermediate errors compound
- Competitive programming problems requiring search
- Scientific reasoning on benchmarks like GPQA Diamond
- Agentic tasks that benefit from planning and reflection
Competitors followed quickly. Google's Gemini 2.0 Flash Thinking, DeepSeek's R1 in early 2025, and Anthropic's extended thinking mode all adopted variants of the paradigm. Some published training recipes openly; others kept them secret.
“We've developed a new series of AI models designed to spend more time thinking before they respond.”
Key terms in this lesson
The big idea: reasoning models reopened the scaling frontier by moving compute from training time into inference time. A model that can think longer is a different kind of model.
End-of-lesson quiz
Check what stuck
15 questions · Score saves to your progress.
Tutor
Curious about “Reasoning Models: OpenAI o1 and After”?
Ask anything about this lesson. I’ll answer using just what you’re reading — short, friendly, grounded.
Progress saved locally in this browser. Sign in to sync across devices.
Related lessons
Keep going
Builders · 25 min
Benchmarks, Leaderboards, and Their Limits
Every new model claims a new high score. Before you trust a leaderboard, learn what benchmarks actually measure — and what they miss.
Builders · 35 min
A Short History: From Expert Systems to Transformers
AI did not start in 2022. It has decades of wrong turns and breakthroughs. Knowing the history helps you spot hype from real progress.
Builders · 22 min
Papers With Code and Reproducibility
A paper without code is often a paper without truth. Papers With Code links claims to runnable proof. Where Claims Meet Code Papers With Code is a community-maintained site that pairs AI papers with their open-source implementations and benchmark results.
