Llama 3.2 Family — Local AI Model by Meta

Meta's multimodal and edge-optimized series. Llama 3.2 introduces vision capabilities (11B and 90B variants) and ultra-compact 1B/3B versions designed to run on smartphones and edge devices. All variants feature a 128k context window.

Hardware Requirements

Llama 3.2 1B InstructMin 1 GB VRAM · Q4_K_M · 128,000 ctx · ollama run llama3.2:1b
Llama 3.2 3B InstructMin 2 GB VRAM · Q4_K_M · 128,000 ctx · ollama run llama3.2:3b
Llama 3.2 11B Vision InstructMin 8 GB VRAM · Q4_K_M · 128,000 ctx · ollama run llama3.2-vision:11b
Llama 3.2 90B Vision InstructMin 48 GB VRAM · Q4_K_M · 128,000 ctx · ollama run llama3.2-vision:90b

How to Run Locally

Install Ollama then run: ollama run llama3.2:1b

Minimum VRAM: 1 GB. For best results use Q4_K_M quantization.