Lesson 224 of 1570
Word2vec: Meaning Becomes Geometry
A 2013 paper from Google showed that words could live as points in space, with analogies as arithmetic.
Lesson map
What this lesson covers
Learning path
The main moves in order
- 1King Minus Man Plus Woman Equals Queen
- 2word2vec
- 3embeddings
- 4distributional semantics
Concept cluster
Terms to connect while reading
Section 1
King Minus Man Plus Woman Equals Queen
In 2013, Tomas Mikolov and colleagues at Google released word2vec, a pair of simple neural network algorithms for learning vector representations of words from raw text. Each word became a point in a few hundred dimensions, learned from the company it kept.
The magic trick that caught everyone's attention: vector arithmetic produced analogies. The vector for king minus man plus woman landed near queen. Paris minus France plus Italy landed near Rome. Structure in language became structure in geometry.
How it learned
- Skip-gram: predict surrounding words from the current word
- CBOW: predict the current word from surrounding context
- Negative sampling: pit true neighbors against random ones to sharpen learning
- Train on billions of words; adjust vectors until predictions improve
Embeddings were not brand new. Older work on latent semantic analysis and neural language models by Bengio had mapped words to vectors before. Word2vec's contribution was speed, scale, and a viral demo that made everyone understand what embeddings were good for.
“You shall know a word by the company it keeps.”
Key terms in this lesson
The big idea: meaning lives in patterns of use, and those patterns can be captured as geometry. This insight powers not just language models but also image models, retrieval, and recommendation systems.
End-of-lesson quiz
Check what stuck
15 questions · Score saves to your progress.
Tutor
Curious about “Word2vec: Meaning Becomes Geometry”?
Ask anything about this lesson. I’ll answer using just what you’re reading — short, friendly, grounded.
Progress saved locally in this browser. Sign in to sync across devices.
Related lessons
Keep going
Builders · 30 min
Tokens and Embeddings: How AI Reads Words
AI does not read letters. It reads tokens, which live as vectors in a space of meaning. Learn how text becomes numbers you can do math on.
Builders · 28 min
Quality Filtering: Separating Signal From Noise
The raw web is 99 percent garbage. Filtering it down to the 1 percent worth training on is one of the highest-leverage steps in modern AI.
Builders · 25 min
Benchmarks, Leaderboards, and Their Limits
Every new model claims a new high score. Before you trust a leaderboard, learn what benchmarks actually measure — and what they miss.
