Zephyr is a series of fine-tuned versions of the Mistral and Mixtral models that are trained to act as helpful assistants.
7b
141b
225.1K Pulls Updated 7 months ago
Updated 10 months ago
10 months ago
887c48b48ef2 · 5.4GB
model
archllama
·
parameters7.24B
·
quantizationQ5_1
5.4GB
params
{"stop":["\u003c|system|\u003e","\u003c|user|\u003e","\u003c|assistant|\u003e","\u003c/s\u003e"]}
98B
template
{{- if .System }}
<|system|>
{{ .System }}
</s>
{{- end }}
<|user|>
{{ .Prompt }}
</s>
<|assistant|>
101B
license
MIT License
Copyright (c) [year] [fullname]
Permission is hereby granted, free of charge, to any p
1.1kB
Readme
Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr 141B-A35B is the latest model in the series, and is a fine-tuned version of Mixtral 8x22b.
Sizes
zephyr:141b
: A Mixture of Experts (MoE) model with 141B total parameters and 35B active parameters.zephyr:7b
: The original Zephyr model