
-
deepseek-r1
DeepSeek-R1 is a family of open reasoning models with performance approaching that of leading models, such as O3 and Gemini 2.5 Pro.
tools thinking 1.5b 7b 8b 14b 32b 70b 671b61.3M Pulls 35 Tags Updated 2 months ago
-
deepseek-v3
A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.
671b2.2M Pulls 5 Tags Updated 8 months ago
-
deepseek-coder
DeepSeek Coder is a capable coding model trained on two trillion code and natural language tokens.
1.3b 6.7b 33b1.2M Pulls 102 Tags Updated 1 year ago
-
deepseek-coder-v2
An open-source Mixture-of-Experts code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks.
16b 236b1.1M Pulls 64 Tags Updated 1 year ago
-
deepscaler
A fine-tuned version of Deepseek-R1-Distilled-Qwen-1.5B that surpasses the performance of OpenAI’s o1-preview with just 1.5B parameters on popular math evaluations.
1.5b304K Pulls 5 Tags Updated 7 months ago
-
deepcoder
DeepCoder is a fully open-Source 14B coder model at O3-mini level, with a 1.5B version also available.
1.5b 14b280.9K Pulls 9 Tags Updated 5 months ago
-
deepseek-llm
An advanced language model crafted with 2 trillion bilingual tokens.
7b 67b202.2K Pulls 64 Tags Updated 1 year ago
-
deepseek-v2
A strong, economical, and efficient Mixture-of-Experts language model.
16b 236b185.3K Pulls 34 Tags Updated 1 year ago
-
exaone-deep
EXAONE Deep exhibits superior capabilities in various reasoning tasks including math and coding benchmarks, ranging from 2.4B to 32B parameters developed and released by LG AI Research.
2.4b 7.8b 32b161.1K Pulls 13 Tags Updated 5 months ago
-
deepseek-v2.5
An upgraded version of DeekSeek-V2 that integrates the general and coding abilities of both DeepSeek-V2-Chat and DeepSeek-Coder-V2-Instruct.
236b66K Pulls 7 Tags Updated 12 months ago
-
deepseek-v3.1
DeepSeek-V3.1 is a hybrid model that supports both thinking mode and non-thinking mode.
tools thinking 671b37.6K Pulls 4 Tags Updated 1 week ago