35 3 weeks ago

My (Gökdeniz Gülmez) first reasoning model fine-tuned on a custom distill dataset

tools thinking
ollama run goekdenizguelmez/Josie-R1:4b-q8_0

Details

3 weeks ago

d979dcaad9c9 · 4.3GB ·

qwen3
·
4.02B
·
Q8_0
{{- $lastUserIdx := -1 -}} {{- range $idx, $msg := .Messages -}} {{- if eq $msg.Role "user" }}{{ $la
Persona: J.O.S.I.E.-R.-1 – Just One Super Intelligent Entity - Reasoning - Version 1, nick named "
Apache License Version 2.0, January 2004 http://www.apache.org/licenses/ TERMS AND CONDITIONS FOR US
{ "min_p": 0, "repeat_penalty": 1, "stop": [ "<|im_start|>", "<|im_end|>

Readme

JOSIE-R1-4B

The JOSIE-R1-4B models are reasoning-oriented language models trained on a custom distillation dataset derived from the original Josie-RL-Zero-V1 model. The training process emphasizes instruction fidelity, reasoning consistency, and practical usefulness across a wide range of everyday and technical tasks.

Model Card for Goekdeniz-Guelmez/JOSIE-R1-4B

Model Description

Introducing JOSIE-R1-4B, a new member of the JOSIE family, fine-tuned with a strong focus on openness, instruction alignment, and transparent reasoning behavior. The model is designed to follow user intent closely while retaining the expressive and occasionally unconventional character that defines the Josie lineage.

Benchmarks comming soon!

  • Developed by: Goekdeniz-Guelmez
  • Funded by: Goekdeniz-Guelmez
  • Shared by: Goekdeniz-Guelmez
  • Model type: qwen3
  • Finetuned from model: Qwen/Qwen3-4B