Qwen3-Coder — Local AI Model by Alibaba Cloud

Alibaba's dedicated coding model released February 2026. The 80B-A3B variant uses a sparse MoE design (only 3B active parameters) yet outperforms DeepSeek V3.2 (37B active) and matches Claude Sonnet 4.5 on SWE-bench Pro. Best-in-class for repo-level coding, debugging, and code review.

Hardware Requirements

Qwen3-Coder 8BMin 6 GB VRAM · Q4_K_M · 128,000 ctx · ollama run qwen3-coder:8b
Qwen3-Coder 80B-A3B (MoE)Min 8 GB VRAM · Q4_K_M · 128,000 ctx · ollama run qwen3-coder:80b-a3b-q4

How to Run Locally

Install Ollama then run: ollama run qwen3-coder:8b

Minimum VRAM: 6 GB. For best results use Q4_K_M quantization.