ollama run batiai/kimi-k2.6:iq3
Frontier 1T MoE from Moonshot AI, quantized directly from official FP8 weights.
| Tag | Size | Min RAM | Target Hardware |
|---|---|---|---|
| q5 | 728GB | 768GB | 2× M3 Ultra 512GB / 8× A100 80GB / H100 node — highest quality |
| iq4 | 546GB | 512GB | M3 Ultra 512GB / 8× A100 80GB / H100 node — recommended |
| iq3 | 394GB | 384GB | M3 Ultra 512GB / H100 node — most accessible |
ollama run batiai/kimi-k2.6:iq4 # recommended balance
ollama run batiai/kimi-k2.6:iq3 # smaller, fits 384GB+ RAM
ollama run batiai/kimi-k2.6:q5 # highest quality, needs 768GB+ RAM
| Your System | IQ3 (394GB) | IQ4 (546GB) | Q5 (728GB) |
|---|---|---|---|
| Mac 16GB | ❌ | ❌ | ❌ |
| Mac 128GB | ❌ | ❌ | ❌ |
| Mac 256GB | ⚠️ heavy swap | ❌ | ❌ |
| Mac 384GB | ⚠️ tight | ❌ | ❌ |
| Mac M3 Ultra 512GB | ✅ | ✅ tight | ❌ |
| 2× M3 Ultra (cluster) | ✅ | ✅ | ✅ |
| 8× A100 80GB | ✅ | ✅ | ✅ |
| H100 node | ✅ fast | ✅ fast | ✅ fast |
This is not a consumer Mac model. For on-device Mac use, see below.
| Your Mac | Recommended |
|---|---|
| 16GB | batiai/gemma4-e4b:q4 |
| 24GB | batiai/gemma4-26b:iq4 |
| 48GB | batiai/qwen3.5-35b:iq4 |
| 96GB | batiai/qwen3.6-35b:iq4 |
| 128GB | batiai/minimax-m2.7:iq3 |
| M3 Ultra 512GB+ | batiai/kimi-k2.6:iq4 (this model) |
general.author=BatiAI / general.url=https://flow.bati.ai signatureBatiFlow is our on-device Mac AI automation app (free, unlimited, local). The smaller models in our lineup (gemma4, qwen3.5-35b, qwen3.6, minimax-m2.7) serve BatiFlow users directly.
Kimi K2.6 is different — it’s a frontier research / workstation model, beyond consumer hardware reach. We quantize it to demonstrate the pipeline handles the full frontier and for researchers / teams with workstation-class GPUs.