Blog
Discord
GitHub
Models
Sign in
Download
Models
Discord
Blog
GitHub
Download
Sign in
mannix
/
mixtral_7bx2_moe
A high-quality Mixture of Experts (MoE) model with open weights by Mistral AI.
A high-quality Mixture of Experts (MoE) model with open weights by Mistral AI.
Cancel
197
Pulls
Updated
7 months ago
q4_K_S
q4_K_S
7.3GB
View all
11 Tags
mixtral_7bx2_moe:q4_K_S
...
/
template
b967f49e09bf · 44B
"[INST] {{ .System }} {{ .Prompt }} [/INST]"