(Unsloth Dynamic Quants) DeepSeek's first-generation of reasoning models with comparable performance to OpenAI-o1, 671B MoE full model, not dense distilled models.
123 Pulls Updated 2 months ago
Updated 2 months ago
2 months ago
08d46664e5ce · 169GB
model
archdeepseek2
·
parameters671B
·
quantizationIQ1_S
169GB
params
{
"stop": [
"<|begin▁of▁sentence|>",
"<|end▁of▁sentence|>",
148B
template
{{- if .System }}{{ .System }}{{ end }}
{{- range $i, $_ := .Messages }}
{{- $last := eq (len (sli
394B
license
MIT License
Copyright (c) 2023 DeepSeek
Permission is hereby granted, free of charge, to any p
1.1kB
Readme
Note: this model requires Ollama 0.5.5 or later.
https://unsloth.ai/blog/deepseekr1-dynamic
https://huggingface.co/unsloth/DeepSeek-R1-GGUF
Quants computing are sponsored by Fujian Xiaowei Technology Co., Ltd. / 12301.cc