1,071 11 months ago

GLM-4-0414 32B with 128k context (YaRN RoPE scaling). Needs ollama 0.6.6

tools
ollama run rhundt/GLM-4-0414-32b-128k-Q4_K_M

Applications

Claude Code
Claude Code ollama launch claude --model rhundt/GLM-4-0414-32b-128k-Q4_K_M
Codex
Codex ollama launch codex --model rhundt/GLM-4-0414-32b-128k-Q4_K_M
OpenCode
OpenCode ollama launch opencode --model rhundt/GLM-4-0414-32b-128k-Q4_K_M
OpenClaw
OpenClaw ollama launch openclaw --model rhundt/GLM-4-0414-32b-128k-Q4_K_M

Models

View all →

Readme

Quantized with YaRN RoPE scaling to 128k context (factor 4). This needs Ollama >=0.6.6 to run. The num_ctx in the Modelfile defaults to 64k just because I don’t have gobs of VRAM.