Gemma 3 — Local AI Model by Google

Google's most capable open model family. Gemma 3 delivers exceptional quality-per-VRAM ratio, with the 4B model outperforming many 7B competitors. Features multimodal capabilities and a 128k context window across all sizes.

Hardware Requirements

Gemma 3 1B InstructMin 1 GB VRAM · Q4_K_M · 32,000 ctx · ollama run gemma3:1b
Gemma 3 4B InstructMin 4 GB VRAM · Q4_K_M · 128,000 ctx · ollama run gemma3:4b
Gemma 3 12B InstructMin 8 GB VRAM · Q4_K_M · 128,000 ctx · ollama run gemma3:12b
Gemma 3 27B InstructMin 16 GB VRAM · Q4_K_M · 128,000 ctx · ollama run gemma3:27b

How to Run Locally

Install Ollama then run: ollama run gemma3:1b

Minimum VRAM: 1 GB. For best results use Q4_K_M quantization.