85 7 months ago

Single file version with (Dynamic Quants) A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.

7 months ago

1150b30b4e73 · 404GB ·

deepseek2
·
671B
·
Q4_K_M
You are a friendly assistant.

Readme

No readme