290 1 month ago

Building upon Mistral Small 3.2 (2506), with added reasoning capabilities, undergoing SFT from Magistral Medium traces and RL on top, it's a small, efficient reasoning model with 24B parameters.(quantized UD-Q5_K_XL)

vision tools thinking 24b

1 month ago

87012da3cbe3 · 18GB ·

llama
·
23.6B
·
Q5_K_M
clip
·
439M
·
F16
{{- range $index, $_ := .Messages }} {{- $last := eq (len (slice $.Messages $index)) 1}} {{- if eq .
You are Magistral Small. When you're not sure about some information or when the user's request requ
{ "min_p": 0, "repeat_penalty": 1, "stop": [ "</s>" ], "temperature": 0.

Readme

Feature Value
vision true (>=0.11.11)
thinking +/-?
tools true
Device Speed, token/s Context VRAM, gb Versions
RTX 3090 24gb ~42 4096 19 UD-Q5_K_XL, 0.12.2
RTX 3090 24gb ~42 15360 21 UD-Q5_K_XL, 0.12.2
M1 Max 32gb ~15 4096 18 UD-Q5_K_XL, 0.12.2
M1 Max 32gb ~15 15360 20 UD-Q5_K_XL, 0.12.2