Lesson 10 of 1570
A Short History: From Expert Systems to Transformers
AI did not start in 2022. It has decades of wrong turns and breakthroughs. Knowing the history helps you spot hype from real progress.
Lesson map
What this lesson covers
Learning path
The main moves in order
- 1The 70-Year Story in Four Acts
- 2AI history
- 3expert systems
- 4deep learning
Concept cluster
Terms to connect while reading
Section 1
The 70-Year Story in Four Acts
AI research began in the 1950s. The field has gone through booms and winters — periods of huge funding followed by collapse. Understanding this rhythm helps you calibrate today's excitement.
Act 1: Symbolic AI (1950s–1980s)
Early researchers believed intelligence was logic. They wrote programs that manipulated symbols according to formal rules. Expert systems of the 1980s, like MYCIN for medical diagnosis, encoded human expert knowledge as if-then rules. They worked for narrow problems but crumbled outside them.
- 1956: Dartmouth Conference coins the term artificial intelligence
- 1966: ELIZA simulates a therapist with simple text patterns
- 1980s: expert systems boom, then bust due to brittleness
Act 2: Statistical ML (1990s–2010)
Researchers pivoted to data-driven methods. Support vector machines, decision trees, and shallow neural networks dominated. IBM's Deep Blue beat Kasparov at chess in 1997, but it was hand-tuned search, not general intelligence.
Act 3: Deep Learning (2012–2017)
In 2012, a neural network called AlexNet won the ImageNet competition by a huge margin, kicking off the deep learning revolution. GPUs, big datasets, and backpropagation combined to finally make deep networks trainable. By 2016, AlphaGo beat the world champion at Go, a feat nobody thought was close.
Act 4: The Transformer Era (2017–today)
The 2017 paper Attention Is All You Need introduced the transformer architecture. It replaced the recurrent networks used for language with a simpler, more parallel structure. Every modern LLM — GPT, Claude, Gemini, Llama — is a transformer at heart.
Compare the options
| Year | Milestone |
|---|---|
| 2017 | Transformer architecture published |
| 2018 | BERT and GPT-1 released |
| 2020 | GPT-3 shows few-shot learning |
| 2022 | ChatGPT makes AI mainstream |
| 2023-2024 | GPT-4, Claude 3, Gemini, open-source Llama |
| 2025-2026 | Reasoning models, multimodality, agentic systems |
“AI winters end not with a new theory, but with enough compute.”
Key terms in this lesson
The big idea: today's AI is the fourth major wave, built on GPUs, internet-scale data, and the transformer. Knowing the cycle helps you see that hype is not new, but neither is real progress.
End-of-lesson quiz
Check what stuck
15 questions · Score saves to your progress.
Tutor
Curious about “A Short History: From Expert Systems to Transformers”?
Ask anything about this lesson. I’ll answer using just what you’re reading — short, friendly, grounded.
Progress saved locally in this browser. Sign in to sync across devices.
Related lessons
Keep going
Creators · 55 min
The Three Ingredients: Data, Compute, Algorithms (Capstone)
Every AI breakthrough of the past decade rests on three interacting ingredients. Synthesize everything you have learned into one working model.
Builders · 25 min
Benchmarks, Leaderboards, and Their Limits
Every new model claims a new high score. Before you trust a leaderboard, learn what benchmarks actually measure — and what they miss.
Builders · 22 min
Papers With Code and Reproducibility
A paper without code is often a paper without truth. Papers With Code links claims to runnable proof. Where Claims Meet Code Papers With Code is a community-maintained site that pairs AI papers with their open-source implementations and benchmark results.
