The latest series of Code-Specific Qwen models, with significant improvements in code generation, code reasoning, and code fixing.

tools 1.5b 7b

272.1K 4 weeks ago

Readme

Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). For Qwen2.5-Coder, three base language models and instruction-tuned language models will be released, 1.5, 7 and 32 (coming soon) billion parameters.

  • Significantly improvements in code generation, code reasoning and code fixing. Base on the strong Qwen2.5, we scale up the training tokens into 5.5 trillion including source code, text-code grounding, Synthetic data, etc.
  • A more comprehensive foundation for real-world applications such as Code Agents. Not only enhancing coding capabilities but also maintaining its strengths in mathematics and general competencies.
  • Long-context Support up to 128K tokens.

References

Blog Post

GitHub

HuggingFace