The Big Five
Meta — Llama 3.2
Sizes: 1B, 3B (text), 11B, 90B (vision)
License: Llama Community License
Strength: Edge deployment, ecosystem
Google — Gemma 3
Sizes: 1B, 4B, 12B, 27B
License: Gemma Terms of Use
Strength: Efficiency, reasoning per param
Microsoft — Phi-4-mini
Sizes: 3.8B
License: MIT
Strength: Math, code, data quality
Alibaba — Qwen 3.5
Sizes: 0.6B, 1.7B, 4B, 9B, 14B, 32B
License: Apache 2.0
Strength: Leaderboard king, multilingual
Mistral — Mistral Small 3.1
Sizes: 24B
License: Apache 2.0
Strength: Best "medium" model
What “Open-Weight” Means
Open-weight means the trained model weights are publicly available for download. You can run, fine-tune, and deploy them. This is different from “open-source” which implies the training data and code are also available.
Why it matters: Open weights let you run the model on your hardware, modify it, and deploy it without API keys or usage limits. Some licenses (like Llama’s) have restrictions for very large commercial use (>700M monthly users), but for most use cases, they’re effectively free.
Key insight: The open-weight model ecosystem has exploded since 2023. You now have multiple high-quality options at every size point from 1B to 30B. Competition between Meta, Google, Microsoft, Alibaba, and Mistral means rapid improvement — each new release pushes the others to do better.