A fine-tuned version of Deepseek-R1-Distilled-Qwen-1.5B that surpasses the performance of OpenAI’s o1-preview with just 1.5B parameters on popular math evaluations.
1.5b
393 Pulls Updated 5 weeks ago
Updated 5 weeks ago
5 weeks ago
752ca9d83330 · 7.1GB
model
archqwen2
·
parameters1.78B
·
quantizationF32
7.1GB
params
{
"stop": [
"<|begin▁of▁sentence|>",
"<|end▁of▁sentence|>",
179B
template
{{- if .System }}{{ .System }}{{ end }}
{{- range $i, $_ := .Messages }}
{{- $last := eq (len (slice
387B
license
MIT License
Copyright (c) 2025 Agentica
Permission is hereby granted, free of charge, to any perso
1.1kB
Readme
This is an uncensored version of agentica-org/DeepScaleR-1.5B-Preview created with abliteration (see remove-refusals-with-transformers to know more about it).
This is a crude, proof-of-concept implementation to remove refusals from an LLM model without using TransformerLens.
References
Donation
Your donation helps us continue our further development and improvement, a cup of coffee can do it.
- bitcoin:
bc1qqnkhuchxw0zqjh2ku3lu4hq45hc6gy84uk70ge