Qwen: Qwen2.5 Coder 7B Instruct
Qwen2.5-Coder-7B-Instruct is a 7B parameter instruction-tuned language model optimized for code-related tasks such as code generation, reasoning, and bug fixing. Based on the Qwen2.5 architecture, it incorporates enhancements like RoPE, SwiGLU, RMSNorm, and GQA attention with support for up to 128K tokens using YaRN-based extrapolation. It is trained on a large corpus of source code, synthetic data, and text-code grounding, providing robust performance across programming languages and agentic coding workflows. This model is part of the Qwen2.5-Coder family and offers strong compatibility with tools like vLLM for efficient deployment. Released under the Apache 2.0 license.
Pricing per 1M Tokens
| Input (Prompt) | $0.03 |
| Output (Completion) | $0.09 |
| Cache Read | Free |
| Cache Write | Free |
| Image | N/A |
Specifications
| Context Length | 33K |
| Max Output Tokens | N/A |
| Input Modalities | Text |
| Output Modalities | Text |
| Tokenizer | Qwen |
| Instruct Type | N/A |
| Top Provider Context | 33K |
| Top Provider Max Output | N/A |
| Moderated | No |
Compare this model
See how Qwen: Qwen2.5 Coder 7B Instruct stacks up against other models.
More from Qwen
Last updated: March 23, 2026
First tracked: March 23, 2026