latest
8.5GB
Llama 3 8B fine-tuned for Lean tactic prediction
8B
52 Pulls Updated 2 months ago
Updated 2 months ago
2 months ago
c7deb0f09aff · 8.5GB
model
archllama
·
parameters8.03B
·
quantizationQ8_0
8.5GB
params
{"stop":["<|start_header_id|>","<|end_header_id|>","<|eot_id|>","\"<|reserved_special_token\""]}
132B
template
<|begin_of_text|><|start_header_id|>user<|end_header_id|>
{{ .Prompt }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
129B
Readme
No readme