Qwen 2.5 Family — Local AI Model by Alibaba Cloud

The coding and math powerhouse of the open-weight world. Qwen 2.5 models consistently outperform Llama 3.1 at equivalent sizes in STEM tasks.

Hardware Requirements

Qwen 2.5 Coder 32BMin 20 GB VRAM · Q4_K_M · 128,000 ctx · ollama run qwen2.5-coder:32b
Qwen 2.5 14B InstructMin 10 GB VRAM · Q4_K_M · 128,000 ctx · ollama run qwen2.5:14b
Qwen 2.5 7B InstructMin 5 GB VRAM · Q4_K_M · 128,000 ctx · ollama run qwen2.5:7b
Qwen 2.5 72B InstructMin 42 GB VRAM · Q4_K_M · 128,000 ctx · ollama run qwen2.5:72b

How to Run Locally

Install Ollama then run: ollama run qwen2.5-coder:32b

Minimum VRAM: 20 GB. For best results use Q4_K_M quantization.