53 2 years ago

This model was built from the https://huggingface.co/paulkl/mixtral-8x7b-instruct-v0.1_Q4b_wa (gguf)

ollama run grf/mixtral_wa_q4_cp

Details

2 years ago

7dc5a45be6b6 · 26GB ·

llama
·
46.7B
·
Q4_0
{ "num_ctx": 32768, "num_thread": 8 }
[INST] {{ .Prompt }} [/INST]

Readme

No readme