Lesson 284 of 2116
Historical Bias: The COMPAS Case Study
Even accurate data can encode an unjust history. The COMPAS recidivism tool shows what happens when AI learns from a biased past.
Lesson map
What this lesson covers
Learning path
The main moves in order
- 1When the Past Is Not Fair
- 2historical bias
- 3COMPAS
- 4recidivism
Concept cluster
Terms to connect while reading
Section 1
When the Past Is Not Fair
COMPAS is a commercial risk-assessment tool used by US courts to predict whether a defendant will re-offend. In 2016, ProPublica published an investigation showing that COMPAS was nearly twice as likely to falsely flag Black defendants as high-risk compared to white defendants. This case became the most famous example of historical bias in AI.
What the numbers showed
Compare the options
| Error type | Black defendants | White defendants |
|---|---|---|
| Falsely labeled high-risk (didn't re-offend) | 45% | 23% |
| Falsely labeled low-risk (did re-offend) | 28% | 48% |
| Overall accuracy | ~65% | ~65% |
Why this happened
COMPAS was trained on historical data: arrests, convictions, re-arrests. But the US criminal justice system has a well-documented history of policing Black neighborhoods more heavily, leading to more arrests for equivalent behavior. The model learned to replicate that historical pattern, not because the programmers intended bias, but because the data itself reflected generations of unequal enforcement.
The fairness impossibility
What we learned
- Predictive accuracy is not fairness
- Historical data encodes historical injustice
- Fairness has multiple, sometimes conflicting definitions
- Audit for disparate impact per group, not just overall accuracy
- High-stakes predictions (liberty, healthcare, loans) demand extra scrutiny
In 2016, Wisconsin's supreme court ruled that COMPAS could be used in sentencing, but only with warnings about its limitations. The debate continues to this day, and COMPAS remains in use in several US jurisdictions.
Key terms in this lesson
The big idea: a model trained on an unjust past will perpetuate that injustice into the future. Technical accuracy is not a defense. AI used in high-stakes decisions demands moral choices, not just statistical ones.
End-of-lesson quiz
Check what stuck
15 questions · Score saves to your progress.
Tutor
Curious about “Historical Bias: The COMPAS Case Study”?
Ask anything about this lesson. I’ll answer using just what you’re reading — short, friendly, grounded.
Progress saved locally in this browser. Sign in to sync across devices.
Related lessons
Keep going
Creators · 30 min
Debate Prep: Researching Both Sides Fast
Debate rewards knowing the other side's best argument better than they do. AI is built for exactly this kind of fast, balanced research.
Creators · 35 min
Running a Literature Review With AI
AI turns weeks of literature review into days — if you know how to use it. Here is a workflow that actually works.
Creators · 30 min
Citing AI-Assisted Work Honestly
The norms for disclosing AI use in research are still being written. Here is the emerging consensus and how to stay on the right side of it.
