The premise
Hallucination is not a bug to be patched — it is intrinsic to how language models work. Mitigation is real and powerful, but elimination is not.
What AI does well here
- Reducing hallucination significantly via RAG and explicit citation requirements
- Asking the model to express uncertainty when it lacks confidence
- Cross-checking generated facts against trusted sources programmatically
- Designing flows that fail gracefully when the model is wrong
What AI cannot do
- Eliminate hallucination entirely
- Reliably distinguish a model's confident-correct from confident-wrong outputs
- Trust the model's own reports of its uncertainty
End-of-lesson check
15 questions · take it digitally for instant feedback at tendril.neural-forge.io/learn/quiz/end-ai-foundations-hallucination-final1-creators
Is hallucination a bug to be patched or intrinsic to language models?
- A simple bug fixable by a patch
- Intrinsic to how language models work
- A network glitch
- A user error only
Which technique meaningfully reduces hallucination?
- Asking 'be honest' once
- Increasing temperature
- Retrieval-augmented generation with explicit citation requirements
- Removing all instructions
What's a useful instruction for uncertain questions?
- Demand confident answers always
- Ban the word 'maybe'
- Forbid citations
- Tell the model to express uncertainty when it lacks confidence
Why cross-check facts programmatically?
- A program can verify against trusted sources at scale
- Programs can read minds
- Programs are always wrong
- Programs are decorative
What does 'fail gracefully' mean in this context?
- Crash the app
- Design flows that recover when the model is wrong
- Hide errors from users
- Always trust the first answer
Can hallucination be eliminated entirely?
- Yes, with the right prompt
- Yes, with a paid plan
- No — only reduced significantly
- Yes, with multilingual prompts
Can a model reliably distinguish its own confident-correct from confident-wrong outputs?
- Yes, models always know
- Only paid models can
- Only frontier models can
- No — that's why external verification matters
Should you trust the model's own reports of its uncertainty?
- Treat them as a hint, not ground truth
- As gospel
- Ignore them entirely
- Only trust them in JSON
For high-stakes outputs (legal, medical, financial), what's the rule?
- Trust freely
- Assume it might be wrong and build verification into the workflow
- Avoid AI entirely
- Print and hope
What concrete experiment compares mitigation techniques?
- Ask once and trust it
- Avoid factual questions
- Ask the same factual question three ways: bare, with sources, with 'say I don't know if uncertain'
- Use only fiction prompts
What is 'grounding' in this context?
- Plugging the model into the wall
- Disabling the model
- Avoiding all output
- Tying output to retrievable, citable sources
Which is true about hallucination across model sizes?
- Even frontier models hallucinate
- Only small models hallucinate
- Frontier models never hallucinate
- All models are equally bad
What is calibration?
- The screen brightness
- How closely a model's expressed confidence tracks reality
- The font size
- The model temperature
What's the danger of 'prompted-away' hallucination?
- It actually solves the problem
- It increases cost
- It feels solved while still happening silently
- It disables the model
What's the right framing for a designer building on top of an LLM?
- Hallucination is a temporary glitch
- Hallucination is the user's fault
- Hallucination is illegal
- Hallucination is a property of the substrate; design around it