Blog
Discord
GitHub
Models
Sign in
Download
Models
Discord
Blog
GitHub
Download
Sign in
mannix
/
mixtral_7bx2_moe
A high-quality Mixture of Experts (MoE) model with open weights by Mistral AI.
A high-quality Mixture of Experts (MoE) model with open weights by Mistral AI.
Cancel
197
Pulls
Updated
7 months ago
q2_K
q2_K
4.8GB
View all
11 Tags
mixtral_7bx2_moe:q2_K
...
/
params
ed11eda7790d · 30B
{
"stop": [
"[INST]",
"[/INST]"
]
}