Lesson 294 of 2116
Distributions: Normal, Power-Law, and Bimodal
Data comes in shapes. The shape determines which tools you can use, and which assumptions will silently betray you.
Lesson map
What this lesson covers
Learning path
The main moves in order
- 1Every Dataset Has a Shape
- 2distributions
- 3normal
- 4power-law
Concept cluster
Terms to connect while reading
Section 1
Every Dataset Has a Shape
Plot any column as a histogram and you will see its distribution. Some distributions are bell-shaped, some are long-tailed, some have two humps. The shape is not cosmetic; it determines what statistical tools work.
The three shapes you must know
Compare the options
| Distribution | Shape | Real example |
|---|---|---|
| Normal (Gaussian) | Symmetric bell | Human heights, measurement errors |
| Power-law | Very long right tail | City populations, YouTube views, wealth |
| Bimodal | Two humps | Commute times (car vs. transit), restaurant sizes |
Normal distributions
Normal distributions show up whenever many small, independent causes add together (Central Limit Theorem). Height is the sum of many genetic and environmental factors, each tiny. Polling errors are the sum of many small deviations. Mean and standard deviation fully describe a normal distribution.
Power-law distributions
Power-laws appear when outcomes multiply rather than add. Rich people can invest and get richer. Popular videos get recommended and get more popular. A tiny fraction of items (the head) dominates everything else (the long tail). In a power-law, mean is essentially meaningless because a handful of extreme values dominate.
Bimodal distributions
A bimodal distribution has two peaks, usually because your data contains two different populations mashed together. Commute times bimodal: car commuters and transit commuters. Restaurant sizes bimodal: independent cafes and chain restaurants. The trick: if you see bimodality, split the data into two populations and analyze each separately.
Always plot your distributions first
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
df = pd.read_csv('some_data.csv')
# Plot histogram to see the shape
for col in df.select_dtypes('number').columns:
df[col].hist(bins=50)
plt.title(col)
plt.show()
# Check for skew
from scipy.stats import skew, kurtosis
print('Skew:', {c: skew(df[c]) for c in df.select_dtypes('number')})
# |skew| > 1 means substantial asymmetryKey terms in this lesson
The big idea: plot first, average later. The shape of your data tells you which tools to trust and which will give you confidently wrong answers.
End-of-lesson quiz
Check what stuck
15 questions · Score saves to your progress.
Tutor
Curious about “Distributions: Normal, Power-Law, and Bimodal”?
Ask anything about this lesson. I’ll answer using just what you’re reading — short, friendly, grounded.
Progress saved locally in this browser. Sign in to sync across devices.
Related lessons
Keep going
Creators · 45 min
Open vs. Closed Models: Philosophy and Strategy
Open-source AI is both a technical movement and a political one. Understand the arguments so you can pick a stack and defend it.
Creators · 32 min
AP Biology: Using AI to Survive the Vocab Tsunami
AP Bio has roughly a thousand terms and four big concepts. NotebookLM and Claude Projects can turn your textbook into a custom tutor that actually knows what you are studying.
Creators · 40 min
Golden-Dataset Curation
A golden dataset is a curated set of hard, representative examples you trust completely. It is the backbone of every serious eval.
