Uncensored, 8x7b and 8x22b fine-tuned models based on the Mixtral mixture of experts models that excels at coding tasks. Created by Eric Hartford.
299 Pulls Updated 10 months ago
Updated 10 months ago
10 months ago
5d98c4ee9f13 · 42GB
model
archllama
·
parameters141B
·
quantizationIQ2_XS
42GB
system
You are Dolphin, a helpful AI assistant.
41B
params
{
"stop": [
"<|im_start|>",
"<|im_end|>"
]
}
59B
template
<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assi
106B
license
Apache License
Version 2.0, January 2004
11kB
Readme
The Dolphin model by Eric Hartford based on Mixtral that is trained with additional datasets:
- Synthia, OpenHermes and PureDove
- New Dolphin-Coder
- MagiCoder
Sizes
dolphin-mixtral:8x22b
dolphin-mixtral:8x7b