64 3 weeks ago

tools thinking
ollama run batiai/qwen3.5-27b:iq4

Applications

Claude Code
Claude Code ollama launch claude --model batiai/qwen3.5-27b:iq4
OpenClaw
OpenClaw ollama launch openclaw --model batiai/qwen3.5-27b:iq4
Hermes Agent
Hermes Agent ollama launch hermes --model batiai/qwen3.5-27b:iq4
Codex
Codex ollama launch codex --model batiai/qwen3.5-27b:iq4
OpenCode
OpenCode ollama launch opencode --model batiai/qwen3.5-27b:iq4

Models

View all →

Readme

Qwen 3.5 27B — Quantized by BatiAI

Quantized from official Alibaba weights. Verified on real Mac hardware.

Models

Tag Size VRAM M4 Max (128GB) Use Case
iq4 14GB 28GB 17.0 t/s 32GB+ Mac

Quick Start

ollama run batiai/qwen3.5-27b:iq4

Why Qwen 3.5 27B?

  • Hybrid architecture (Gated DeltaNet + GQA + MoE)
  • 262K context window
  • Excellent Korean + tool calling + coding
  • Apache 2.0 license

RAM Requirements

Your Mac RAM IQ4 (14GB)
16GB
32GB ✅ Fits (28GB VRAM)
48GB+ ✅ Fast
128GB 17.0 t/s

Comparison

Model Size VRAM Speed (M4 Max)
batiai/qwen3.5-9b:q4 5.2GB 12.5 t/s (16GB Mac)
batiai/qwen3.5-27b:iq4 14GB 28GB 17.0 t/s
batiai/qwen3.5-35b:iq4 17GB 23GB 26.6 t/s

For 32GB Mac, 27B is the best Qwen option. For 36GB+, the 35B MoE is faster.

Why BatiAI?

  • Quantized directly from official Alibaba weights
  • IQ4_XS with imatrix — best quality at this size
  • Verified on MacBook Pro M4 Max (128GB)
  • Korean, tool calling, JSON generation all tested

Built for BatiFlow

Free, on-device AI automation for Mac. 5MB app, 100% local, unlimited.

https://flow.bati.ai