846 4 months ago

GLM-4-0414 32B with 128k context (YaRN RoPE scaling). Needs ollama 0.6.6

tools

4 months ago

67a2a027906e · 20GB

glm4
·
32.6B
·
Q4_K_M
[gMASK]<sop>{{- /* ---------- tools section ---------- */}} {{- if .Tools }} <|system|> # Available
{ "num_ctx": 64000, "stop": [ "<|system|>", "<|user|>", "<|assistant

Readme

Quantized with YaRN RoPE scaling to 128k context (factor 4). This needs Ollama >=0.6.6 to run. The num_ctx in the Modelfile defaults to 64k just because I don’t have gobs of VRAM.