1 Download Updated 5 hours ago
This is a merge of pre-trained language models created using mergekit.
This model is a revision of the ertghiu256/Qwen3-4b-tcomanr-merge-v2.2
This model aims to combine the reasoning, code, and math capabilities of Qwen3 4b 2507 reasoning by merging it with some other Qwen3 finetunes. This model reasoning is very long.
Run this command
ollama run ertghiu256/Qwen3-4b-tcomanr-merge-v2.3:Q8_0
or for Q5_K_M quant
ollama run hf.co/ertghiu256/Qwen3-4b-tcomanr-merge-v2.3:Q5_K_M
or for IQ4_NL quant
ollama run hf.co/ertghiu256/Qwen3-4b-tcomanr-merge-v2.3:IQ4_NL
Look the huggingface page
temp: 0.6
num_ctx: ≥8192
top_p: 0.95
top_k: 10
Repeat Penalty: 1.1
This model was merged using the TIES merge method using Qwen/Qwen3-4B-Thinking-2507 as a base.
The following models were included in the merge: * ertghiu256/qwen-3-4b-mixture-of-thought * Tesslate/UIGEN-T3-4B-Preview-MAX * ertghiu256/Qwen3-4B-Thinking-2507-Hermes-3 * ValiantLabs/Qwen3-4B-ShiningValiant3 * ertghiu256/qwen3-math-reasoner * ValiantLabs/Qwen3-4B-Esper3 * Qwen/Qwen3-4b-Instruct-2507 * ertghiu256/qwen3-multi-reasoner * janhq/Jan-v1-4B * ertghiu256/qwen3-4b-code-reasoning * ertghiu256/Qwen3-Hermes-4b * GetSoloTech/Qwen3-Code-Reasoning-4B * POLARIS-Project/Polaris-4B-Preview * huihui-ai/Huihui-Qwen3-4B-Thinking-2507-abliterated
The following YAML configuration was used to produce this model:
models:
- model: ertghiu256/qwen3-math-reasoner
parameters:
weight: 0.8
- model: ertghiu256/qwen3-4b-code-reasoning
parameters:
weight: 0.9
- model: ertghiu256/qwen-3-4b-mixture-of-thought
parameters:
weight: 1.0
- model: POLARIS-Project/Polaris-4B-Preview
parameters:
weight: 0.8
- model: ertghiu256/qwen3-multi-reasoner
parameters:
weight: 0.9
- model: ertghiu256/Qwen3-Hermes-4b
parameters:
weight: 0.7
- model: ValiantLabs/Qwen3-4B-Esper3
parameters:
weight: 0.75
- model: Tesslate/UIGEN-T3-4B-Preview-MAX
parameters:
weight: 1.0
- model: ValiantLabs/Qwen3-4B-ShiningValiant3
parameters:
weight: 0.6
density: 0.5
- model: huihui-ai/Huihui-Qwen3-4B-Thinking-2507-abliterated
parameters:
weight: 0.75
- model: Qwen/Qwen3-4B-Thinking-2507
parameters:
weight: 1.0
- model: Qwen/Qwen3-4b-Instruct-2507
parameters:
weight: 0.75
- model: GetSoloTech/Qwen3-Code-Reasoning-4B
parameters:
weight: 0.75
density: 0.55
- model: ertghiu256/Qwen3-4B-Thinking-2507-Hermes-3
parameters:
weight: 1.0
- model: janhq/Jan-v1-4B
parameters:
weight: 0.3
merge_method: ties
base_model: Qwen/Qwen3-4B-Thinking-2507
parameters:
normalize: true
int8_mask: true
lambda: 1.0
dtype: float16