Lesson 100 of 2244
Reproducibility: Making Your AI-Assisted Work Re-Runnable
AI-assisted research is especially vulnerable to reproducibility failures. Model versions shift, prompts drift, outputs vary. Here's how to lock it down.
Adults & Professionals · Research & Analysis · ~6 min read · Interactive
Why AI-assisted work fails reproducibility tests
Traditional reproducibility says: publish your code and data, and anyone can re-run. AI adds three new failure modes: the model you used may be deprecated; the same prompt may produce different outputs next week; and stochastic sampling means even identical prompts produce different outputs across runs.
The lock-down checklist
- 1Record the exact model ID (e.g., 'gpt-4o-2024-08-06', 'claude-sonnet-4-5-20250929')
- 2Record the temperature, top-p, and max-tokens settings
- 3Set a random seed where the API supports it
- 4Version-control every prompt as a text file, not buried in a notebook
- 5Cache the raw model outputs alongside the derived results
- 6Document the date of every API call — model behavior changes silently
What to publish
- The full prompts (as .txt or .md files in your repo)
- The model IDs and settings (as a config file)
- The raw responses you actually used (as JSON)
- Any post-processing code
- A README that walks through re-running the whole pipeline
💡 Try it yourself
Open an AI assistant (Claude, ChatGPT, or Gemini) and practice what you just learned.
→ Apply the concept from this section in a real prompt
→ Note what worked and what surprised you
Key terms in this lesson
The big idea: AI-assisted work is only reproducible if you treat prompts as code and cache outputs as data. Everything else is a promise you can't keep.
End-of-lesson quiz
Check what stuck
13 questions · Score saves to your progress.
Tutor
Curious about “Reproducibility: Making Your AI-Assisted Work Re-Runnable”?
Ask anything about this lesson. I’ll answer using just what you’re reading — short, friendly, grounded.
Progress saved locally in this browser. Sign in to sync across devices.
Related lessons
Keep going
Adults & Professionals · 10 min
Quantitative Analysis Prompting: Asking For Reproducible Code
When you ask an LLM to 'analyze this data,' you get a guess. When you ask it to write reproducible code, you get a collaborator.
Adults & Professionals · 40 min
Literature Review With LLMs: Scope First, Search Second
Use an LLM to define the scope of your lit review before touching a search engine — the single highest-leverage move in modern research workflow.
Adults & Professionals · 40 min
Grant Writing Assistance: Specific Aims, Specifically
Grant writing rewards structural discipline. AI is a near-perfect drafting partner — if you feed it the right scaffolds.
