Qwen2.5-72B-Instruct
The Qwen2.5 series was Alibaba's broadest open release: seven sizes from 0.5B to 72B, all under the same architecture and trained on a substantially expanded 18-trillion-token corpus. The 72B-Instruct is the flagship and one of the strongest non-reasoning open-weight models in the catalog.
Strengths
Where Llama optimizes for general English-language usefulness, Qwen2.5-72B is notably stronger on multilingual tasks (29+ languages), structured output (JSON, tool calls), and Chinese-language work. It also has unusually clean coding behavior — the Qwen2.5-Coder variants are derived from the same base.
Variants worth knowing about
The 7B and 14B Instruct checkpoints are popular as fine-tuning bases because they're small enough to train cheaply and large enough to be genuinely useful. Qwen2.5-Coder-32B-Instruct is widely regarded as the best open-weight coding model under 100B params.
Running it locally
Similar story to Llama 3.3 70B: ~145 GB full precision, ~42 GB at Q4_K_M GGUF. Runs comfortably on a single 48 GB GPU at usable speed or split across two 24 GB cards.
License
Released under the Qwen License, which is permissive for most commercial use but has a service-scale clause for products with over 100 million monthly active users. The smaller Qwen2.5 variants (0.5B through 14B) are released under Apache 2.0.