Zephyr is a series of fine-tuned versions of the Mistral and Mixtral models that are trained to act as helpful assistants.
7b
141b
236.4K Pulls Updated 10 months ago
Updated 14 months ago
14 months ago
bbe38b81adec · 4.1GB
model
archllama
·
parameters7.24B
·
quantizationQ4_0
4.1GB
params
{
"stop": [
"<|system|>",
"<|user|>",
"<|assistant|>",
"</s>"
98B
template
{{- if .System }}
<|system|>
{{ .System }}
</s>
{{- end }}
<|user|>
{{ .Prompt }}
</s>
<|assistant|>
101B
license
MIT License
Copyright (c) [year] [fullname]
Permission is hereby granted, free of charge, to any p
1.1kB
Readme
Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr 141B-A35B is the latest model in the series, and is a fine-tuned version of Mixtral 8x22b.
Sizes
zephyr:141b
: A Mixture of Experts (MoE) model with 141B total parameters and 35B active parameters.zephyr:7b
: The original Zephyr model