Blog
Discord
GitHub
Models
Sign in
Download
Models
Discord
Blog
GitHub
Download
Sign in
mannix
/
mixtral_7bx2_moe
A high-quality Mixture of Experts (MoE) model with open weights by Mistral AI.
A high-quality Mixture of Experts (MoE) model with open weights by Mistral AI.
Cancel
197
Pulls
Updated
7 months ago
q6_K
q6_K
11GB
View all
11 Tags
mixtral_7bx2_moe:q6_K
...
/
params
ed11eda7790d · 30B
{
"stop": [
"[INST]",
"[/INST]"
]
}