18 7 months ago

A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.

7 months ago

74b583cd0ec8 · 404GB

deepseek2
·
671B
·
Q4_K_M
DEEPSEEK LICENSE AGREEMENT Version 1.0, 23 October 2023 Copyright (c) 2023 DeepSeek Section I: PREAM
{ "num_ctx": 64000, "stop": [ "<|begin▁of▁sentence|>", "<|end▁of
{{- range $i, $_ := .Messages }} {{- if eq .Role "user" }}<|User|> {{- else if eq .Role "assista

Readme

Note: this model requires Ollama 0.5.5 or later.

deepseek

DeepSeek-V3 achieves a significant breakthrough in inference speed over previous models. It tops the leaderboard among open-source models and rivals the most advanced closed-source models globally.

This is a modified version of deepseek-v3 Q4_K_M with context length set to 64k