A compact, yet powerful 10.7B large language model designed for single-turn conversation.
10.7b
62.2K Pulls Updated 11 months ago
Updated 11 months ago
11 months ago
7e7394f2f53e · 6.1GB
model
archllama
·
parameters10.7B
·
quantizationQ4_K_S
6.1GB
template
### System:
{{ .System }}
### User:
{{ .Prompt }}
### Assistant:
67B
params
{
"num_ctx": 4096,
"stop": [
"</s>",
"### System:",
"### User:",
86B
Readme
Solar is the first open-source 10.7 billion parameter language model. It’s compact, yet remarkably powerful, and demonstrates state-of-the-art performance in models with parameters under 30B.
This model leverages the Llama 2 architecture and employs the Depth Up-Scaling technique, integrating Mistral 7B weights into upscaled layers.
On the H6 benchmark, this model outperforms models with up to 30B parameters, even the Mixtral 8X7B model.