Llama 3.3 70B Instruct
Meta's flagship dense instruction-tuned model. Matches the original 405B on most benchmarks at a fraction of the inference cost.
Fifteen of the most useful open-weight language models in one place — with honest specs, license notes, and direct download links to the original publishers. No login wall, no marketing fluff.
Meta's flagship dense instruction-tuned model. Matches the original 405B on most benchmarks at a fraction of the inference cost.
671B-parameter Mixture-of-Experts with 37B active per token. Strong on code and math; competitive with the top closed models at far lower training cost.
A reasoning model trained with reinforcement learning. Exposes a transparent chain-of-thought and is unusually good at multi-step problems.
A general-purpose dense model that punches above its weight in multilingual and coding tasks. The 7B/14B/32B variants are popular for fine-tuning.
Sparse Mixture-of-Experts with 8 experts of 22B parameters each. Fast for its size thanks to only 39B active params per token.
The 7B that made local inference real for a lot of people. Still a sensible baseline when you want something fast on a single GPU.
Google's open model family. The 27B is the sweet spot: noticeably stronger than the 9B and still fits on a single high-end consumer GPU at 4-bit.
A 14B model trained heavily on synthetic data. Punches well above its weight on reasoning and math benchmarks for its size.
104B dense model tuned for retrieval-augmented generation and tool use. Among the best open-weights options for grounded enterprise workflows.
A bilingual (Chinese/English) model with strong long-context retrieval. The 34B is a popular choice for serious self-hosting on a 48GB card.
A compact bilingual chat model with surprisingly strong instruction-following. Competitive with much larger Western models in its size class.
TII's third-generation Falcon. The 10B targets edge deployment with respectable English-only performance and a very permissive license.
A genuinely open model: weights, training data, and code all published. The natural pick when reproducibility and provenance matter.
A genuinely small model that's still useful. Designed for on-device deployment where 8B is too heavy.
IBM's enterprise-leaning open model. Stable, well-documented, and oriented toward business workflows rather than benchmark chasing.