latest
14GB
7B
18 Pulls Updated 5 months ago
Updated 5 months ago
5 months ago
7fdddd2adf2c · 14GB
model
archllama
·
parameters7.24B
·
quantizationF16
14GB
system
You are an AI assistant. Be as helpful as possible.
53B
template
<|im_start|>
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assistant
101B
params
{"stop":["<|im_end|>","<|im_start|>"]}
59B
Readme
This is a non-quantized version of Dolphin 2.8-Mistral which was created used the ollama create
command.