CodeQwen1.5 is a large language model pretrained on a large amount of code data.
7b
112.1K Pulls Updated 4 months ago
Updated 4 months ago
4 months ago
df352abf55b1 · 4.2GB
model
archqwen2
·
parameters7.25B
·
quantizationQ4_0
4.2GB
system
You are a helpful assistant.
28B
license
../../../Qwen/CodeQwen1.5-7B-Chat/LICENSE
41B
params
{"stop":["\u003c|im_start|\u003e","\u003c|im_end|\u003e"]}
59B
template
{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
182B
Readme
CodeQwen1.5 is based on Qwen1.5. It is trained on 3 trillion tokens of code data. Its major features include:
- Strong code generation capabilities and competitive performance across a series of benchmarks
- Support for long context understanding and generation with a maximum context length of 64K tokens
- Support for 92 coding languages
- Excellent performance in Text-to-SQL, fixing bugs and other coding use cases.