Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
Hanversion
/
MAGA-T1-Tieba-1.5B-Distill
:latest
47
Downloads
Updated
6 months ago
This is a distill model that trained from the dataset of TieBa latest. Used about 8k data and think chain from DeepSeek-V3.
This is a distill model that trained from the dataset of TieBa latest. Used about 8k data and think chain from DeepSeek-V3.
Cancel
Updated 6 months ago
6 months ago
f6f38d9ed892 · 1.9GB ·
model
arch
qwen2
·
parameters
1.78B
·
quantization
Q8_0
1.9GB
template
<|begin▁of▁sentence|>{{ if .System }}{{ .System }}{{ end }}{{ range .Messages }}{{ if eq .Ro
248B
params
{ "num_ctx": 4096, "stop": [ "<|end▁of▁sentence|>" ] }
66B
Readme
No readme
Write
Preview
Paste, drop or click to upload images (.png, .jpeg, .jpg, .svg, .gif)