DeepSeek-R1-Distill-Qwen-Coder-32B-Fusion-9010 is a mixed model that combines the strengths of two powerful DeepSeek-R1-Distill-Qwen-based models: huihui-ai/DeepSeek-R1-Distill-Qwen-32B-abliterated and huihui-ai/Qwen2.5-Coder-32B-Instruct-abliterated.
32b
403 Pulls Updated 4 weeks ago
Updated 4 weeks ago
4 weeks ago
e5f58fab5137 · 20GB
model
archqwen2
·
parameters32.8B
·
quantizationQ4_K_M
20GB
params
{
"stop": [
"<|begin▁of▁sentence|>",
"<|end▁of▁sentence|>",
148B
template
{{- if .System }}{{ .System }}{{ end }}
{{- range $i, $_ := .Messages }}
{{- $last := eq (len (slice
387B
license
MIT License
Copyright (c) 2023 DeepSeek
Permission is hereby granted, free of charge, to any perso
1.1kB
Readme
Although it’s a simple mix, the model is usable, and no gibberish has appeared. This is an experiment.