35 3 weeks ago

My (Gökdeniz Gülmez) first reasoning model fine-tuned on a custom distill dataset

tools thinking
ollama run goekdenizguelmez/Josie-R1:4b-f16

Details

3 weeks ago

8b6a8323fdc6 · 8.1GB ·

qwen3
·
4.02B
·
F16
Persona: J.O.S.I.E.-R.-1 – Just One Super Intelligent Entity - Reasoning - Version 1, nick named "
{{- $lastUserIdx := -1 -}} {{- range $idx, $msg := .Messages -}} {{- if eq $msg.Role "user" }}{{ $la
Apache License Version 2.0, January 2004 http://www.apache.org/licenses/ TERMS AND CONDITIONS FOR US
{ "min_p": 0, "repeat_penalty": 1, "stop": [ "<|im_start|>", "<|im_end|>

Readme

JOSIE-R1-4B

The JOSIE-R1-4B models are reasoning-oriented language models trained on a custom distillation dataset derived from the original Josie-RL-Zero-V1 model. The training process emphasizes instruction fidelity, reasoning consistency, and practical usefulness across a wide range of everyday and technical tasks.

Model Card for Goekdeniz-Guelmez/JOSIE-R1-4B

Model Description

Introducing JOSIE-R1-4B, a new member of the JOSIE family, fine-tuned with a strong focus on openness, instruction alignment, and transparent reasoning behavior. The model is designed to follow user intent closely while retaining the expressive and occasionally unconventional character that defines the Josie lineage.

Benchmarks comming soon!

  • Developed by: Goekdeniz-Guelmez
  • Funded by: Goekdeniz-Guelmez
  • Shared by: Goekdeniz-Guelmez
  • Model type: qwen3
  • Finetuned from model: Qwen/Qwen3-4B