Gemma 4 — Local AI Model by Google

Google's fourth-generation open model family, released in early 2026. Gemma 4 significantly improves reasoning, coding, and multilingual capability over Gemma 3 while maintaining the same hardware-friendly VRAM footprint. The 1B and 4B variants run on phones and budget GPUs; the 27B competes with 70B-class models.

Hardware Requirements

Gemma 4 1B InstructMin 1 GB VRAM · Q4_K_M · 32,000 ctx · ollama run gemma4:1b
Gemma 4 4B InstructMin 4 GB VRAM · Q4_K_M · 128,000 ctx · ollama run gemma4:4b
Gemma 4 12B InstructMin 8 GB VRAM · Q4_K_M · 128,000 ctx · ollama run gemma4:12b
Gemma 4 27B InstructMin 16 GB VRAM · Q4_K_M · 256,000 ctx · ollama run gemma4:27b

How to Run Locally

Install Ollama then run: ollama run gemma4:1b

Minimum VRAM: 1 GB. For best results use Q4_K_M quantization.