Gemma 3 27B
Gemma 3 27B is the largest Gemma 3 model — a dense 27B with vision support, 128K context, and strong multilingual coverage. At Q4_K_M it needs ~15 GB VRAM, fitting on 16 GB GPUs with limited context or 24 GB with room to spare. The GeGLU ac…
27.0B
Parameters
128K
Max Context
Dense
Architecture
Mar 12, 2025
Released
Text + Vision
Modality
About Gemma 3 27B
Gemma 3 27B is the largest Gemma 3 model — a dense 27B with vision support, 128K context, and strong multilingual coverage. At Q4_K_M it needs ~15 GB VRAM, fitting on 16 GB GPUs with limited context or 24 GB with room to spare. The GeGLU activation and unique head_dim=176 give it a distinctive performance profile. Strong across general knowledge, coding, and multilingual tasks. A solid alternative to Qwen 2.5 32B for users who prefer Google's training methodology.
Technical Specifications
System Requirements
Estimated VRAM at 10% overhead for different quantization methods and context sizes.
| Quantization | 1K ctx | 128K ctx |
|---|---|---|
Q4_K_M0.50 B/W ~97% of FP16 | 14.30Consumer GPU | 57.96Datacenter GPU |
Q8_01.00 B/W ~100% of FP16 | 28.26Datacenter GPU | 71.91Datacenter GPU |
F162.00 B/W Reference | 56.17Datacenter GPU | 99.82Cluster / Multi-GPU |
Other Gemma Models
View AllFind the right GPU for Gemma 3 27B
Use the interactive VRAM Calculator to see exactly how much memory you need at any quantization level, context length, and overhead setting.