Gemma 4 27B
MoEGemma
Google's 27B MoE model with only 4B active parameters per token. Near-frontier quality at a fraction of compute cost.
Provider
Parameters
27B (4B active MoE)
Context
128K
Released
2026-04-08
VRAM Requirements by Quantization
| Method | Disk Size | VRAM Required | Fits GPUs |
|---|---|---|---|
| Q8_0 | 28 GB | 30 GB | 3 GPUs |
| Q4_K_M | 15.5 GB | 17 GB | 7 GPUs |
| Q4_0 | 14.8 GB | 16 GB | 7 GPUs |
| Q2_K | 10 GB | 11.5 GB | 12 GPUs |
Install with Ollama
Benchmark Scores
mmlu83.5%
humaneval80.1%
Scores are approximate and may vary by quantization level.
Compatible GPUs (12)
HuggingFace
google/gemma-4-27b-it