Loading lesson…
The AI safety ecosystem is small, influential, and often misunderstood. Here is who does what, how they get funded, and how to tell real work from rhetoric.
When people say AI safety research, they can mean dozens of different groups doing very different things. Some do empirical evals. Some do interpretability. Some do policy. Some do field-building. Pretending they all agree or do the same thing is a common mistake.
| Actor | Access to model internals | Independent of lab | Can stop a release |
|---|---|---|---|
| Internal safety team | Yes | No | Sometimes (company governance) |
| METR/Apollo | API or sandboxed | Yes | No (but publish findings) |
| UK/US AISI | Pre-release, under NDA | Yes (government) | No formal veto yet |
| EU AI Office | Documentation, testing rights | Yes | Yes for systemic-risk GPAI |
| Academic researchers | Mostly public API | Yes | No |
The 2024 Seoul AI Safety Summit produced the Frontier AI Safety Commitments, where 16 major labs pledged specific pre-deployment evals and capability thresholds. The 2025 Paris Summit rebranded as the AI Action Summit and broadened focus to economic AI. Governance is bifurcating: pre-deployment safety evals (AISI-style) on one track, general AI policy (AI Act, Executive Orders) on another.
Safety is not a department. It is a property of the whole system, and it emerges from the culture as much as the team.
— Helen Toner, former OpenAI board member
The big idea: AI safety is an actual ecosystem with real people doing real work. Knowing the map — who does what, who funds what, who can stop what — lets you read any AI safety headline with the context it deserves.
15 questions · take it digitally for instant feedback at tendril.neural-forge.io/learn/quiz/end-ethics-safety-orgs-creators
What is the core idea behind "AI Safety Orgs and How They Actually Operate"?
Which term best describes a foundational idea in "AI Safety Orgs and How They Actually Operate"?
A learner studying AI Safety Orgs and How They Actually Operate would need to understand which concept?
Which of these is directly relevant to AI Safety Orgs and How They Actually Operate?
Which of the following is a key point about AI Safety Orgs and How They Actually Operate?
Which of these does NOT belong in a discussion of AI Safety Orgs and How They Actually Operate?
Which statement is accurate regarding AI Safety Orgs and How They Actually Operate?
Which of these does NOT belong in a discussion of AI Safety Orgs and How They Actually Operate?
What is the key insight about "How to tell real work from rhetoric" in the context of AI Safety Orgs and How They Actually Operate?
What is the recommended tip about "Key insight" in the context of AI Safety Orgs and How They Actually Operate?
Which statement accurately describes an aspect of AI Safety Orgs and How They Actually Operate?
What does working with AI Safety Orgs and How They Actually Operate typically involve?
Which of the following is true about AI Safety Orgs and How They Actually Operate?
Which best describes the scope of "AI Safety Orgs and How They Actually Operate"?
Which section heading best belongs in a lesson about AI Safety Orgs and How They Actually Operate?