Qwen2.5 models are pretrained on Alibaba's latest large-scale dataset, encompassing up to 18 trillion tokens. The model supports up to 128K tokens and has multilingual support.
tools
0.5b
1.5b
3b
7b
14b
32b
72b
1.9M Pulls Updated 7 weeks ago
Updated 7 weeks ago
7 weeks ago
6ed3506191d1 · 4.1GB
model
archqwen2
·
parameters7.62B
·
quantizationQ3_K_L
4.1GB
system
You are Qwen, created by Alibaba Cloud. You are a helpful assistant.
68B
template
{{- if .Messages }}
{{- if or .System .Tools }}<|im_start|>system
{{- if .System }}
{{ .System }}
{{
1.5kB
license
Apache License
Version 2.0, January 200
11kB
Readme
Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, a range of base language models and instruction-tuned models are released, with sizes ranging from 0.5 to 72 billion parameters. Qwen2.5 introduces the following improvements over Qwen2:
- It possesses significantly more knowledge and has greatly enhanced capabilities in coding and mathematics, due to specialized expert models in these domains.
- It demonstrates significant advancements in instruction following, long-text generation (over 8K tokens), understanding structured data (e.g., tables), and generating structured outputs, especially in JSON format. It is also more resilient to diverse system prompts, improving role-play and condition-setting for chatbots.
- It supports long contexts of up to 128K tokens and can generate up to 8K tokens.
- It offers multilingual support for over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more.
Please note: all models except the 3B and 72B are released under the Apache 2.0 license, while the 3B and 72B models are under the Qwen license.