1 5 hours ago

This model aims to combine the reasoning, code, and math capabilities of Qwen3 4b 2507 reasoning by merging it with some other Qwen3 finetunes.

tools thinking

Models

View all →

Readme

Ties merged COde MAth aNd Reasoning model

This is a merge of pre-trained language models created using mergekit.

Merge Details

This model is a revision of the ertghiu256/Qwen3-4b-tcomanr-merge-v2.2

This model aims to combine the reasoning, code, and math capabilities of Qwen3 4b 2507 reasoning by merging it with some other Qwen3 finetunes. This model reasoning is very long.

How to run

Ollama

Run this command

ollama run ertghiu256/Qwen3-4b-tcomanr-merge-v2.3:Q8_0

or for Q5_K_M quant

ollama run hf.co/ertghiu256/Qwen3-4b-tcomanr-merge-v2.3:Q5_K_M

or for IQ4_NL quant

ollama run hf.co/ertghiu256/Qwen3-4b-tcomanr-merge-v2.3:IQ4_NL

Others

Look the huggingface page

Recomended parameters

temp: 0.6
num_ctx: ≥8192
top_p: 0.95
top_k: 10
Repeat Penalty: 1.1

Merge Method

This model was merged using the TIES merge method using Qwen/Qwen3-4B-Thinking-2507 as a base.

Models Merged

The following models were included in the merge: * ertghiu256/qwen-3-4b-mixture-of-thought * Tesslate/UIGEN-T3-4B-Preview-MAX * ertghiu256/Qwen3-4B-Thinking-2507-Hermes-3 * ValiantLabs/Qwen3-4B-ShiningValiant3 * ertghiu256/qwen3-math-reasoner * ValiantLabs/Qwen3-4B-Esper3 * Qwen/Qwen3-4b-Instruct-2507 * ertghiu256/qwen3-multi-reasoner * janhq/Jan-v1-4B * ertghiu256/qwen3-4b-code-reasoning * ertghiu256/Qwen3-Hermes-4b * GetSoloTech/Qwen3-Code-Reasoning-4B * POLARIS-Project/Polaris-4B-Preview * huihui-ai/Huihui-Qwen3-4B-Thinking-2507-abliterated

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: ertghiu256/qwen3-math-reasoner
    parameters:
      weight: 0.8
  - model: ertghiu256/qwen3-4b-code-reasoning
    parameters:
      weight: 0.9
  - model: ertghiu256/qwen-3-4b-mixture-of-thought
    parameters:
      weight: 1.0
  - model: POLARIS-Project/Polaris-4B-Preview
    parameters:
      weight: 0.8
  - model: ertghiu256/qwen3-multi-reasoner
    parameters:
      weight: 0.9
  - model: ertghiu256/Qwen3-Hermes-4b
    parameters:
      weight: 0.7
  - model: ValiantLabs/Qwen3-4B-Esper3
    parameters:
      weight: 0.75
  - model: Tesslate/UIGEN-T3-4B-Preview-MAX
    parameters:
      weight: 1.0
  - model: ValiantLabs/Qwen3-4B-ShiningValiant3
    parameters:
      weight: 0.6
      density: 0.5
  - model: huihui-ai/Huihui-Qwen3-4B-Thinking-2507-abliterated
    parameters:
      weight: 0.75
  - model: Qwen/Qwen3-4B-Thinking-2507
    parameters:
      weight: 1.0
  - model: Qwen/Qwen3-4b-Instruct-2507
    parameters:
      weight: 0.75
  - model: GetSoloTech/Qwen3-Code-Reasoning-4B
    parameters:
      weight: 0.75
      density: 0.55
  - model: ertghiu256/Qwen3-4B-Thinking-2507-Hermes-3
    parameters:
      weight: 1.0
  - model: janhq/Jan-v1-4B
    parameters:
      weight: 0.3
merge_method: ties
base_model: Qwen/Qwen3-4B-Thinking-2507
parameters:
  normalize: true
  int8_mask: true
  lambda: 1.0
dtype: float16