Loading lesson…
Asking the model to critique and revise its own output is one of the cheapest quality boosts in prompt engineering. Master the patterns and their limits.
Generating text and evaluating text are slightly different cognitive operations. When a model is asked to critique an answer (especially without being told it's its own answer), it can spot flaws it missed when generating. This is the foundation behind Constitutional AI and many production pipelines.
Task: Write a short explanation of why inflation happens, aimed at a 9th-grade audience.
Step 1: Draft your answer in <draft> tags.
Step 2: In <critique> tags, evaluate the draft:
- Is it accurate? Flag any oversimplifications.
- Is it level-appropriate? Point out any jargon.
- Is it complete? Name one thing missing.
Step 3: In <final> tags, rewrite the draft fixing the issues raised.
Only the <final> block will be shown to the student.A single-prompt self-critique — all three stages in one call.This works because Claude will genuinely evaluate its own draft when asked. The <final> block is almost always better than the <draft>, often markedly so. The cost is extra tokens (1.5-2x) for significantly higher quality.
TURN 1:
Write a resignation letter. (no mention of critique)
TURN 2:
You are a no-nonsense HR consultant. The letter below was written by an employee. Identify the five biggest problems and rewrite it.
<letter>
{RESPONSE_FROM_TURN_1}
</letter>The AI has no idea it's critiquing itself. This sometimes produces harsher, better critique than self-aware critique.A related technique: generate N answers at temperature > 0, then use the model (or majority vote) to pick the best. On reasoning tasks, 5 samples with majority vote on the final answer dramatically outperforms a single sample.
# Pseudocode
answers = []
for i in range(5):
answer = model.complete(prompt, temperature=0.7)
answers.append(extract_final_answer(answer))
best = majority_vote(answers) # or ask a judge modelSelf-consistency with majority voting.15 questions · take it digitally for instant feedback at tendril.neural-forge.io/learn/quiz/end-prompting-self-critique-creators
What is the core idea behind "Self-Critique Prompts: AI as Its Own Reviewer"?
Which term best describes a foundational idea in "Self-Critique Prompts: AI as Its Own Reviewer"?
A learner studying Self-Critique Prompts: AI as Its Own Reviewer would need to understand which concept?
Which of these is directly relevant to Self-Critique Prompts: AI as Its Own Reviewer?
Which of the following is a key point about Self-Critique Prompts: AI as Its Own Reviewer?
Which of these does NOT belong in a discussion of Self-Critique Prompts: AI as Its Own Reviewer?
What is the key insight about "Don't stack critique endlessly" in the context of Self-Critique Prompts: AI as Its Own Reviewer?
What is the key insight about "Related reading" in the context of Self-Critique Prompts: AI as Its Own Reviewer?
What is the recommended tip about "Practitioner tip" in the context of Self-Critique Prompts: AI as Its Own Reviewer?
Which statement accurately describes an aspect of Self-Critique Prompts: AI as Its Own Reviewer?
What does working with Self-Critique Prompts: AI as Its Own Reviewer typically involve?
Which of the following is true about Self-Critique Prompts: AI as Its Own Reviewer?
Which best describes the scope of "Self-Critique Prompts: AI as Its Own Reviewer"?
Which section heading best belongs in a lesson about Self-Critique Prompts: AI as Its Own Reviewer?
Which section heading best belongs in a lesson about Self-Critique Prompts: AI as Its Own Reviewer?