The latest series of Code-Specific Qwen models, with significant improvements in code generation, code reasoning, and code fixing.
tools
1.5b
7b
272.1K Pulls Updated 4 weeks ago
Updated 4 weeks ago
4 weeks ago
05df1f695342 · 940MB
model
archqwen2
·
parameters1.54B
·
quantizationQ4_K_S
940MB
params
{"stop":["\u003c|endoftext|\u003e"]}
37B
system
You are Qwen, created by Alibaba Cloud. You are a helpful assistant.
68B
template
{{- if .Suffix }}<|fim_prefix|>{{ .Prompt }}<|fim_suffix|>{{ .Suffix }}<|fim_middle|>
{{- else if .M
1.6kB
license
Apache License
Version 2.0, January 200
11kB
Readme
Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). For Qwen2.5-Coder, three base language models and instruction-tuned language models will be released, 1.5, 7 and 32 (coming soon) billion parameters.
- Significantly improvements in code generation, code reasoning and code fixing. Base on the strong Qwen2.5, we scale up the training tokens into 5.5 trillion including source code, text-code grounding, Synthetic data, etc.
- A more comprehensive foundation for real-world applications such as Code Agents. Not only enhancing coding capabilities but also maintaining its strengths in mathematics and general competencies.
- Long-context Support up to 128K tokens.