Loading lesson…
There are too many open-weight models. A short, opinionated tour of the major families and what each is actually good at.
Open-weight models cluster into a few families with distinct personalities and strengths. Within each family, sizes range from 1B (laptop-friendly) to 70B+ (small cluster). Knowing the families saves you from drowning in Hugging Face.
| Family | Origin | Sweet spot | Reputation |
|---|---|---|---|
| Llama | Meta | General purpose, broad ecosystem | The default — well-supported |
| Mistral / Mixtral | Mistral AI (France) | Efficient, strong reasoning per parameter | European, MoE-friendly |
| Qwen | Alibaba | Coding, multilingual, long context | Often best-in-class at small sizes |
| DeepSeek | DeepSeek (China) | Reasoning and coding | Punches well above its size |
| Hermes / Nous | Community fine-tunes | Chattier, less refusal-y | Fine-tunes of base models |
| Phi | Microsoft Research | Tiny but capable | Great for embedded / edge |
| Gemma | Light, well-tuned | Polished, conservative |
The big idea: the right local model is the one that wins on your prompts, on your hardware. Family names are a starting filter, not an answer.
15 questions · take it digitally for instant feedback at tendril.neural-forge.io/learn/quiz/end-local-choosing-a-model-creators
What is the core idea behind "Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends"?
Which term best describes a foundational idea in "Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends"?
A learner studying Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends would need to understand which concept?
Which of these is directly relevant to Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?
Which of the following is a key point about Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?
Which of these does NOT belong in a discussion of Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?
Which statement is accurate regarding Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?
Which of these does NOT belong in a discussion of Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?
What is the key insight about "Base vs Instruct vs Chat" in the context of Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?
What is the key insight about "Fine-tunes are a wild west" in the context of Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?
What is the key insight about "From the community" in the context of Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?
Which statement accurately describes an aspect of Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?
What does working with Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends typically involve?
Which best describes the scope of "Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends"?
Which section heading best belongs in a lesson about Choosing a Local Model: Llama, Mistral, Hermes, Qwen, DeepSeek, and Friends?