Qwen2.5 Coder 32B with the corrected 128k context
tools
333 Pulls Updated 2 months ago
Updated 2 months ago
2 months ago
bf89d7131e9e · 27GB
model
archqwen2
·
parameters32.8B
·
quantizationQ6_K
27GB
params
{
"min_p": 0.9,
"num_ctx": 32768,
"num_keep": 256,
"repeat_penalty": 1.05,
"stop
130B
system
You are Qwen, created by Alibaba Cloud. You are a helpful assistant.
You are an expert programmer.
538B
template
{{- if .Suffix }}<|fim_prefix|>{{ .Prompt }}<|fim_suffix|>{{ .Suffix }}<|fim_middle|>
{{- else if .
1.6kB
Readme
This uses Unsloth’s GGUF which fixes the context length (the official Ollama model is wrong).