finetuning by alpaca_chinese_dataset
9 Pulls Updated 2 months ago
Updated 2 months ago
2 months ago
66aa5ab3c443 · 4.9GB
model
archllama
·
parameters8.03B
·
quantizationQ4_K_M
4.9GB
params
{"stop":["\u003c|end|\u003e","\u003c|user|\u003e","\u003c|assistant|\u003e"]}
78B
template
{{ if .System }}<|system|>
{{ .System }}<|end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|end
149B
Readme
finetuning by alpaca_chinese_dataset
Maximum number of tokens:16K