189 Downloads Updated yesterday
ollama run mdq100/qwen3.5-coder:122b
Updated yesterday
yesterday
ad2feedda0f2 · 81GB ·

Coding-optimized variants of the official Qwen3.5 MoE models — full vision capability retained, tuned for precise code generation via lower temperature. Based on Alibaba’s Qwen3.5 distributed through the Ollama registry.
Two tags available:
| Tag | Parameters | Active/Token | Context | Vision |
|---|---|---|---|---|
mdq100/qwen3.5-coder:35b |
36.0B | ~3B | 262,144 | Yes |
mdq100/qwen3.5-coder:122b |
125.1B | ~3B | 262,144 | Yes |
Same architecture. Same quantization. Coding-optimized temperature across both.
Qwen3.5 is a hybrid Mixture-of-Experts model from Alibaba’s Qwen team featuring a novel Gated DeltaNet + sparse MoE architecture. Despite large total parameter counts, only ~3B are activated per token, making inference efficient regardless of model size.
These Coder variants take the official Ollama Qwen3.5 models and tune their parameters for coding workloads — lower temperature for more deterministic output, presence penalty to reduce repetition. Vision capability is fully preserved.
OpenCode and similar coding tools don’t support per-session parameter overrides — they use whatever is baked into the Ollama model. These variants provide coding-optimized defaults without sacrificing any model capability.
:35b for fast, capable coding on standard hardware:122b for maximum reasoning depth on high-memory systems| Property | :35b |
:122b |
|---|---|---|
| Architecture | qwen35moe | qwen35moe |
| Total parameters | 36.0B | 125.1B |
| Active per token | ~3B | ~3B |
| Context length | 262,144 | 262,144 |
| Embedding length | 2048 | 3072 |
| Quantization | Q4_K_M | Q4_K_M |
| Source | Ollama registry | Ollama registry |
The 122b model has a larger embedding dimension (3072 vs 2048), enabling richer per-token representations and stronger reasoning depth for complex problems.
ollama pull mdq100/qwen3.5-coder:35b
ollama run mdq100/qwen3.5-coder:35b
Parameters:
temperature: 0.6
top_p: 0.95
top_k: 20
ollama pull mdq100/qwen3.5-coder:122b
ollama run mdq100/qwen3.5-coder:122b
Parameters:
temperature: 0.6
top_p: 0.95
top_k: 20
num_ctx: 131072
:35b):{
"model": "ollama/mdq100/qwen3.5-coder:35b"
}
:122b for complex tasks):{
"model": "ollama/mdq100/qwen3.5-coder:122b"
}
Scores from the base Qwen3.5 model (BF16, full precision). Q4_K_M quantization may show minor variance (~1-2%).
| Benchmark | Score |
|---|---|
| SWE-bench Verified | 69.2 |
| LiveCodeBench v6 | 74.6 |
| CodeForces Rating | 2028 |
| FullStackBench (en) | 58.1 |
| Terminal Bench 2 | 40.5 |
| Benchmark | Score |
|---|---|
| MMLU-Pro | 85.3 |
| MMLU-Redux | 93.3 |
| GPQA Diamond | 84.2 |
| HLE w/ CoT | 22.4 |
| Benchmark | Score |
|---|---|
| IFEval | 91.9 |
| IFBench | 70.2 |
| MultiChallenge | 60.0 |
| Benchmark | Score |
|---|---|
| LongBench v2 | 59.0 |
| AA-LCR | 58.5 |
| Benchmark | Score |
|---|---|
| MMMU | 81.4 |
| MathVision | 83.9 |
| MMBenchEN-DEV-v1.1 | 91.5 |
| Property | mdq100/qwen3.5-coder |
mdq100/qwen3.5-flash:35b-code |
|---|---|---|
| Source | Ollama registry (official) | Unsloth GGUF |
| Vision | Yes | No |
| Quantization | Standard Q4_K_M | Unsloth Dynamic 2.0 |
| Ollama validated | Yes | No |
| Tags | :35b, :122b |
:35b |
Use mdq100/qwen3.5-coder when you need vision support or prefer Ollama-validated builds.
Use mdq100/qwen3.5-flash:35b-code for better quantization quality without vision.