7,708 Downloads Updated 1 month ago
ollama run fredrezones55/Qwopus3.5:9b-Q8_0
My internet is a tad slow, the models might take a bit to upload… Will upload other quants of the model later.
An attempt to create a direct port of the Qwopus3.5 -v3 GGUF format Finetuned family of optimized reasoning models with the Ollama engine. [the conversion was much faster as I did not have to debug my pipeline, same base model as Qwen3.5-Opus series]. Core changes made was to port the model to how the Ollama engine would want it. The finetune was created by: https://huggingface.co/Jackrong
Qwopus3.5-v3 is a reasoning-enhanced model based on Qwen3.5. Its core objective is to simultaneously improve reasoning stability and correctness while optimizing inference efficiency, ultimately achieving stronger cross-task generalization capabilities—particularly in programming.
While the overall accuracy margin (+1.43 pp) is modest, Qwopus3.5-9B-v3 fundamentally shifts the accuracy-cost paradigm, achieving its victory while spending significantly less reasoning budget. With a 25.3% reduction in mean think length and 24.0% lower token cost per correct answer, this iteration is highly optimized for latency, token budget, and context pressure.
Furthermore, across the mixed domain profile, Qwopus3.5-9B-v3 uniquely offsets Qwen3.5-9B’s slight edge in biology, CS, and math by excelling in physics, chemistry, and significantly lowering its unfinished-output rate. Its final rank benefits as much from raw correctness as from an improved ability to cleanly and reliably complete analytical boundaries.
Base Model (Qwen3.5-9B)
│
▼
Qwen3.5-9B fine-tuned with Unsloth
│
▼
Supervised Fine-Tuning (SFT) + LoRA
(Response-Only Training masked on "<|im_start|>assistant\n<think>")
│
▼
Qwopus3.5-9B-v3
The model includes targeted optimizations addressing Qwen3.5’s tendency toward excessive or repetitive reasoning on simple queries. By distilling the structured reasoning habits of top-tier models like Claude Opus, Qwopus3.5-v3 adopts a highly organized, step-by-step cognitive layout.
Example:The user is asking about [Topic] and how it differs from [Topic B]. This is a [Task type] question. Let me break this down:
1. What is [Topic A]?
- [Fact/Mechanism 1]
- [Fact/Mechanism 2]
2. What is [Topic B]?
- [Fact/Mechanism 1]
3. Key differences:
- [Comparison Point 1]
- [Comparison Point 2]
Let me make sure to be accurate: [...]
Actually, I should double-check: is [Fact] used before [Fact]? Yes, typically...
Let me provide a clear, well-structured answer:
The model was fine-tuned on a high-fidelity reasoning dataset, which was meticulously curated from a blend of premium open-source sources on Hugging Face. This dataset is the result of a rigorous mixing and cleaning process, specifically designed to filter out low-quality responses and ensure consistently strong logical performance across diverse analytical domains.
(Rest assured, the entire process is strictly by-the-book and 100% compliant with all terms and open-source licenses!)