latest
4.9GB
finetuning by alpaca_chinese_dataset
8B
3 Pulls Updated 2 weeks ago
Updated 3 weeks ago
3 weeks ago
66aa5ab3c443 · 4.9GB
model
archllama
·
parameters8.03B
·
quantizationQ4_K_M
4.9GB
template
{{ if .System }}<|system|>
{{ .System }}<|end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|end|>
{{ end }}<|assistant|>
{{ .Response }}<|end|>
149B
params
{"stop":["<|end|>","<|user|>","<|assistant|>"]}
78B
Readme
finetuning by alpaca_chinese_dataset
Maximum number of tokens:16K