-
qwen3-4b-thinking-2507_q8
Qwen3-4B-Thinking-2507_q8: A 4-billion-parameter inference model with 8-bit quantization, optimized for efficient reasoning in resource-constrained environments.
tools thinking409 Pulls 1 Tag Updated 3 months ago
-
qwen3-4b_q8
A 4B-parameter, 8-bit quantized inference model with /think (reasoning) and /no_think (fast response) modes.
tools thinking141 Pulls 1 Tag Updated 3 months ago