DeepSeek Coder is a capable coding model trained on two trillion code and natural language tokens.
1.4M Pulls 102 Tags Updated 1 year ago
An open-source Mixture-of-Experts code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks.
1.1M Pulls 64 Tags Updated 1 year ago
An upgraded version of DeekSeek-V2 that integrates the general and coding abilities of both DeepSeek-V2-Chat and DeepSeek-Coder-V2-Instruct.
68.4K Pulls 7 Tags Updated 1 year ago
3,299 Pulls 23 Tags Updated 1 year ago
A strong, economical, and efficient Mixture-of-Experts language model with Tool Calling support.
3,007 Pulls 3 Tags Updated 8 months ago
Fine-tuned version of deepseek-ai/deepseek-coder-1.3b-base using 0.5B of TypeScript code
2,560 Pulls 3 Tags Updated 1 year ago
https://huggingface.co/bartowski/DeepSeek-Coder-V2-Lite-Instruct-GGUF
2,375 Pulls 22 Tags Updated 1 year ago
1,146 Pulls 1 Tag Updated 1 year ago
Quantized version of DeepSeek Coder v1.5 and Q8_0_L quantization of v2 model form bartowski/DeepSeek-Coder-V2-Lite-Base-GGUF and bartowski/DeepSeek-Coder-V2-Lite-Instruct-GGUF
845 Pulls 14 Tags Updated 1 year ago
This model was converted to GGUF format from deepseek-ai/DeepSeek-Coder-V2-Lite-Base using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model.
563 Pulls 1 Tag Updated 10 months ago
189 Pulls 1 Tag Updated 1 year ago
170 Pulls 3 Tags Updated 1 year ago
166 Pulls 1 Tag Updated 1 year ago
This is a brand new Mixture of Export (MoE) model from DeepSeek, specializing in coding instructions. (quantized IQ4_XS)
158 Pulls 1 Tag Updated 3 weeks ago
134 Pulls 1 Tag Updated 1 year ago
122 Pulls 1 Tag Updated 1 year ago
With correct 16k context window
106 Pulls 5 Tags Updated 1 year ago
102 Pulls 1 Tag Updated 1 year ago
91 Pulls 1 Tag Updated 3 months ago
59 Pulls 1 Tag Updated 1 year ago