Qwen 2.5 Coder 7B
Qwen 2.5 Coder 7B is the most popular local coding model in the 7B class. Fine-tuned from Qwen 2.5 7B on a massive code corpus spanning 92 programming languages, it delivers exceptional code completion, bug fixing, and code explanation. At …
7.6B
Parameters
32K
Max Context
Dense
Architecture
Nov 12, 2024
Released
Text
Modality
About Qwen 2.5 Coder 7B
Qwen 2.5 Coder 7B is the most popular local coding model in the 7B class. Fine-tuned from Qwen 2.5 7B on a massive code corpus spanning 92 programming languages, it delivers exceptional code completion, bug fixing, and code explanation. At ~4 GB VRAM at Q4_K_M it runs on any GPU. Mature GGUF and MLX support. The Apache 2.0 license makes it safe for commercial development workflows. If you can only run one local coding model, this is the default recommendation.
Technical Specifications
System Requirements
Estimated VRAM at 10% overhead for different quantization methods and context sizes.
| Quantization | 1K ctx | 32K ctx |
|---|---|---|
Q4_K_M0.50 B/W ~97% of FP16 | 3.99Consumer GPU | 5.68Consumer GPU |
Q8_01.00 B/W ~100% of FP16 | 7.92Consumer GPU | 9.62Consumer GPU |
F162.00 B/W Reference | 15.79Consumer GPU | 17.48Consumer GPU |
Other Qwen Models
View AllFind the right GPU for Qwen 2.5 Coder 7B
Use the interactive VRAM Calculator to see exactly how much memory you need at any quantization level, context length, and overhead setting.