BitNet b1.58 — Local AI Model by Microsoft

A research paradigm shift by Microsoft. BitNet b1.58 replaces 16-bit weights with ternary {-1, 0, 1} weights. This eliminates Matrix Multiplication (MatMul) in favor of simple addition, offering extreme speed and efficiency on CPUs.

Hardware Requirements

BitNet b1.58 3BMin 2 GB VRAM · 1.58-bit · 2,048 ctx · ollama run hf.co/1bitLLM/bitnet_b1_58-3B

How to Run Locally

Install Ollama then run: ollama run hf.co/1bitLLM/bitnet_b1_58-3B

Minimum VRAM: 2 GB. For best results use Q4_K_M quantization.