A strong, economical, and efficient Mixture-of-Experts language model with Tool Calling.
tools
16b
236b
170 Pulls Updated 2 months ago
Updated 2 months ago
2 months ago
cbd4d9d19ddd · 133GB
model
archdeepseek2
·
parameters236B
·
quantizationQ4_0
133GB
params
{"min_p":0,"mirostat":0,"mirostat_eta":0.1,"mirostat_tau":5,"num_ctx":4096,"num_predict":128,"repeat
224B
template
{{- if .Messages }}
{{- if or .System .Tools }}
{{- if .System }}
{{ .System }}
{{- end }}
{{
1.3kB
license
DEEPSEEK LICENSE AGREEMENT
Version 1.0, 23 October 2023
Copyright (c) 2023 DeepSeek
Section I: PR
14kB
Readme
No readme