Lesson 642 of 2116
LoRA and Fine-Tuning: When Prompting Is Not Enough
Students should know when to prompt, when to use RAG, and when a small adapter or fine-tune is actually justified.
Lesson map
What this lesson covers
Learning path
The main moves in order
- 1The operational idea: local fine-tuning decisions
- 2LoRA
- 3fine-tuning
- 4adapter
Concept cluster
Terms to connect while reading
Section 1
The operational idea: local fine-tuning decisions
Students should know when to prompt, when to use RAG, and when a small adapter or fine-tune is actually justified. In local AI, the model family is only one part of the system. The runtime, file format, serving path, hardware budget, evaluation set, and safety policy decide whether the model becomes useful.
Compare the options
| Layer | What to decide | What can go wrong |
|---|---|---|
| Runtime | local fine-tuning decisions | The model runs, but the workflow is slow or brittle |
| Evaluation | A small task-specific test set | A flashy demo hides routine failures |
| Safety and ops | Permissions, provenance, logging, and rollback | Fine-tuning on a tiny, messy dataset and making the model worse while believing it became specialized. |
Current source signal
Build the small version
Make a decision tree that chooses prompting, RAG, LoRA, or full fine-tuning for different failure modes.
- 1Define the user task in one sentence.
- 2Choose the smallest model and runtime that might pass that task.
- 3Run one happy-path prompt and one failure-path prompt.
- 4Record speed, memory pressure, output quality, and the exact reason for any failure.
- 5Write the operating rule you would give a non-expert user.
A local-model operations sketch students can adapt.
adaptation_decision:
if model_lacks_current_facts: use_RAG
if output_style_is_wrong: improve_prompt_or_examples
if repeated_format_task_with_many_examples: consider_LoRA
if broad_capability_gap: choose_better_base_model
never: tune_without_eval_setKey terms in this lesson
The big idea: tune only with evals. A local model app is not done when the model answers once; it is done when the whole workflow can be installed, measured, trusted, and recovered.
End-of-lesson quiz
Check what stuck
15 questions · Score saves to your progress.
Tutor
Curious about “LoRA and Fine-Tuning: When Prompting Is Not Enough”?
Ask anything about this lesson. I’ll answer using just what you’re reading — short, friendly, grounded.
Progress saved locally in this browser. Sign in to sync across devices.
Related lessons
Keep going
Creators · 10 min
Fine-Tuning Hermes For A Specific Domain
Fine-tuning a model that is already a fine-tune sounds redundant. It is not. Hermes is a strong starting point precisely because the second-pass tune does less heavy lifting.
Creators · 40 min
When to Fine-Tune vs When to Just Prompt: A Decision Framework
Fine-tuning is expensive and slow to iterate on. Prompting is fast and free. Knowing when fine-tuning actually pays off saves teams from premature optimization.
Builders · 8 min
When Fine-Tuning Actually Beats Just Writing a Better Prompt
Fine-tune for style and format consistency at high volume; for everything else, prompt better first.
