Loading lesson…
Training a frontier model uses the electricity of a small city for months. Running inference at scale matches a large country's load. Here is what the numbers actually look like.
Every token your favorite AI generates came out of a GPU that was drawing electricity from a grid that was, somewhere upstream, burning something or spinning something. The abstraction is clean. The physics is not.
Training a model takes weeks or months and then stops. Running it serves billions of users every day for years. By 2025, most analyses estimated that inference consumed more energy than training across the industry. The IEA projected that global data center electricity use could reach 945 TWh by 2030, with AI as the fastest-growing slice.
Data centers are cooled by enormous amounts of water. A 2023 study estimated GPT-3 training consumed about 700,000 liters of fresh water. Microsoft's water use jumped 34 percent from 2021 to 2022, partly attributed to AI. In drought-prone regions like Arizona, this has become a political fight.
| Activity | Rough energy cost |
|---|---|
| Google search | ~0.3 Wh |
| LLM chat turn | ~3-10 Wh |
| Image generation | ~30-100 Wh |
| Video generation (per second) | ~300-1000 Wh |
| Train GPT-3 once | ~1.3 GWh |
| Train frontier 2025 model | 100+ GWh |
You cannot compute your way out of thermodynamics. Somebody always pays the electric bill.
— A data center engineer
The big idea: AI has real physical costs that are often hidden behind a clean chat interface. Whether they are worth it is a judgment call — but you cannot make the call without knowing the numbers.
15 questions · take it digitally for instant feedback at tendril.neural-forge.io/learn/quiz/end-ethics-environmental-cost-builders
What is the core idea behind "The Environmental Cost of Training a Big Model"?
Which term best describes a foundational idea in "The Environmental Cost of Training a Big Model"?
A learner studying The Environmental Cost of Training a Big Model would need to understand which concept?
Which of these is directly relevant to The Environmental Cost of Training a Big Model?
Which of the following is a key point about The Environmental Cost of Training a Big Model?
Which of these does NOT belong in a discussion of The Environmental Cost of Training a Big Model?
Which statement is accurate regarding The Environmental Cost of Training a Big Model?
Which of these does NOT belong in a discussion of The Environmental Cost of Training a Big Model?
What is the key insight about "Per-query numbers" in the context of The Environmental Cost of Training a Big Model?
What is the key insight about "Reporting quality is poor" in the context of The Environmental Cost of Training a Big Model?
What is the recommended tip about "Key insight" in the context of The Environmental Cost of Training a Big Model?
Which statement accurately describes an aspect of The Environmental Cost of Training a Big Model?
What does working with The Environmental Cost of Training a Big Model typically involve?
Which of the following is true about The Environmental Cost of Training a Big Model?
Which best describes the scope of "The Environmental Cost of Training a Big Model"?