Gemma 4
modelFour-variant family released under Apache 2.0 (a notable license upgrade from prior Gemma releases). Includes a 31B dense flagship, 26B-A4B MoE (4B active), and edge variants E4B and E2B. 256K token context for the larger models. Multimodal: text, image, video, audio.
Architecture features alternating attention, dual RoPE, Per-Layer Embeddings (PLE), and shared KV cache. Gemma 4 31B achieves 85.2% MMLU Pro, 89.2% AIME 2026, 80.0% LiveCodeBench v6, and Codeforces Elo 2,150. Ranked #3 open model on Arena AI text leaderboard. Crossed 2M downloads shortly after release.
Model Details
Architecture DENSE
Parameters 31B
Context window 256,000
Variants
| Name | Parameters | Notes |
|---|---|---|
| Gemma 4 31B | 31B | Dense flagship, 256K context |
| Gemma 4 26B-A4B | 26B | MoE, 256K context |
| Gemma 4 E4B | 8B | Effective 4.5B, 128K context |
| Gemma 4 E2B | 5.1B | Effective 2.3B, 128K context |