latest
1.6GB
https://huggingface.co/notmebug/Qwen2-1.5B-fourchannel
1.5B
2 Pulls Updated 4 weeks ago
Updated 4 weeks ago
4 weeks ago
11252db23711 · 1.6GB
model
archqwen2
·
parameters1.54B
·
quantizationQ8_0
1.6GB
params
{"stop":["<|im_start|>","<|im_end|>"]}
59B
template
{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ .Response }}<|im_end|>
182B
Readme
No readme