LLM Configurator — Free GPU VRAM Checker for Local AI Models

LLM Configurator is the definitive free tool for checking GPU compatibility with local LLMs. Enter your GPU's VRAM and system RAM to instantly discover which open-source AI models you can run — with Ollama install commands, speed estimates, and electricity cost calculations.

VRAM Requirements Quick Reference

VRAMModels You Can Run
2–4 GBSmolLM2 1.7B, Phi-3.5 Mini, BitNet 3B, Gemma 3 4B (Q4), Llama 3.2 1B/3B
6–8 GBLlama 3.1 8B, Gemma 3 4B (FP16), Qwen 2.5 7B, Phi-4 Mini, DeepSeek R1 8B
8–12 GBPhi-4 14B (Q4), Qwen 2.5 14B (Q4), Mistral NeMo 12B, Gemma 3 12B (Q4)
12–16 GBLlama 4 Scout 17B (Q4), Qwen 3 14B, Qwen 3 30B-A3B (MoE)
16–24 GBQwen 3 32B (Q4), Mistral Small 3.1 24B (Q4), Gemma 3 27B (Q4)
24+ GBLlama 3.3 70B (Q4), DeepSeek R1 32B, Llama 4 Maverick (Q4)

Featured Models

Llama 4 (Meta, 2025)

Scout: 17B active / 109B total (MoE). Requires ~10 GB VRAM at Q4. ollama run llama4:scout. Maverick: 17B active / 400B total. Requires ~24 GB VRAM. ollama run llama4:maverick

DeepSeek R1

State-of-the-art reasoning model in 8B, 14B, 32B, and 671B variants. The 8B distill requires 6 GB VRAM: ollama run deepseek-r1:8b

Gemma 3 (Google, 2025)

Available in 1B (2 GB), 4B (3 GB), 12B (8 GB Q4), and 27B (18 GB Q4). ollama run gemma3:4b

Qwen 3 (Alibaba, 2025)

8B, 14B, 32B dense + MoE variants 30B-A3B (fits 6 GB!) and 235B-A22B. ollama run qwen3:8b

Mistral Small 3.1

24B multimodal (vision + text). 16 GB VRAM at Q4. ollama run mistral-small3.1

Setup Guides

About LLM Configurator

Free, independent tool for the local AI community. Supports 75+ open-source models. No account required. No ads. Free forever. Contact: contact@llmconfigurator.com