3,046 6 days ago

9B coding agent based on Qwen3.5-9B, fine-tuned on 425K real agentic traces from Claude Opus 4.6, GPT-5.4, and Gemini 3.1. Reads before it writes, traces bugs to the root cause, doesn't clobber your existing code.

tools thinking
ollama run carstenuhlig/omnicoder-9b

Applications

Claude Code
Claude Code ollama launch claude --model carstenuhlig/omnicoder-9b
Codex
Codex ollama launch codex --model carstenuhlig/omnicoder-9b
OpenCode
OpenCode ollama launch opencode --model carstenuhlig/omnicoder-9b
OpenClaw
OpenClaw ollama launch openclaw --model carstenuhlig/omnicoder-9b

Models

View all →

Readme

Qwen3.5-9B fine-tuned on 425K agentic coding trajectories from Claude Opus 4.6, GPT-5.4, and Gemini 3.1 Pro. Trained on what frontier agents actually do when editing code: read before writing, trace errors to the root cause, write diffs not rewrites.

Weaker on niche languages (Haskell, MATLAB, assembly) and general knowledge — the training data skews heavily toward Python/JS.

Variants

Tag Size
latest / q4_k_m 5.7 GB
q8_0 9.5 GB

Community

From the r/LocalLLaMA thread:

“The single biggest failure mode we hit with smaller models in agentic loops is they just start writing code without checking what’s already there. Ends up clobbering imports, duplicating functions, the usual mess.”

Test on 2× RTX 5060 Ti (Q8): matched a 30B MoE on a FastAPI refactoring task. OmniCoder handled async/sync correctly and produced a clean single diff. The 30B duplicated the diff block and mixed up AsyncSession with sync Session. Prompt eval: 3076 tok/s.

For agentic use, lower temperature to 0.2–0.4. Benchmarks and training details: huggingface.co/Tesslate/OmniCoder-9B