742 6 months ago

DeepSeek's first-generation of reasoning models with comparable performance to OpenAI-o1, including six dense models distilled from DeepSeek-R1 based on Llama and Qwen.

65481175e980 · 46GB
    Metadata
  • split.count
    5
  • split.no
    2
  • split.tensors.count
    1025
  • Tensor
    blk.26
  • blk.26.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.26.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.26.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.26.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.26.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.26.ffn_norm.weight
    F32
    [7168]
  • blk.26.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.26.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.27
  • blk.27.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.27.attn_kv_a_norm.weight
    F32
    [512]
  • blk.27.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.27.attn_norm.weight
    F32
    [7168]
  • blk.27.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.27.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.27.attn_q_a_norm.weight
    F32
    [1536]
  • blk.27.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.27.exp_probs_b.bias
    F32
    [256]
  • blk.27.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.27.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.27.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.27.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.27.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.27.ffn_norm.weight
    F32
    [7168]
  • blk.27.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.27.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.28
  • blk.28.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.28.attn_kv_a_norm.weight
    F32
    [512]
  • blk.28.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.28.attn_norm.weight
    F32
    [7168]
  • blk.28.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.28.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.28.attn_q_a_norm.weight
    F32
    [1536]
  • blk.28.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.28.exp_probs_b.bias
    F32
    [256]
  • blk.28.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.28.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.28.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.28.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.28.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.28.ffn_norm.weight
    F32
    [7168]
  • blk.28.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.28.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.29
  • blk.29.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.29.attn_kv_a_norm.weight
    F32
    [512]
  • blk.29.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.29.attn_norm.weight
    F32
    [7168]
  • blk.29.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.29.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.29.attn_q_a_norm.weight
    F32
    [1536]
  • blk.29.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.29.exp_probs_b.bias
    F32
    [256]
  • blk.29.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.29.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.29.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.29.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.29.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.29.ffn_norm.weight
    F32
    [7168]
  • blk.29.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.29.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.30
  • blk.30.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.30.attn_kv_a_norm.weight
    F32
    [512]
  • blk.30.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.30.attn_norm.weight
    F32
    [7168]
  • blk.30.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.30.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.30.attn_q_a_norm.weight
    F32
    [1536]
  • blk.30.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.30.exp_probs_b.bias
    F32
    [256]
  • blk.30.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.30.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.30.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.30.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.30.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.30.ffn_norm.weight
    F32
    [7168]
  • blk.30.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.30.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.31
  • blk.31.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.31.attn_kv_a_norm.weight
    F32
    [512]
  • blk.31.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.31.attn_norm.weight
    F32
    [7168]
  • blk.31.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.31.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.31.attn_q_a_norm.weight
    F32
    [1536]
  • blk.31.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.31.exp_probs_b.bias
    F32
    [256]
  • blk.31.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.31.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.31.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.31.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.31.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.31.ffn_norm.weight
    F32
    [7168]
  • blk.31.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.31.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.32
  • blk.32.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.32.attn_kv_a_norm.weight
    F32
    [512]
  • blk.32.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.32.attn_norm.weight
    F32
    [7168]
  • blk.32.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.32.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.32.attn_q_a_norm.weight
    F32
    [1536]
  • blk.32.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.32.exp_probs_b.bias
    F32
    [256]
  • blk.32.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.32.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.32.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.32.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.32.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.32.ffn_norm.weight
    F32
    [7168]
  • blk.32.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.32.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.33
  • blk.33.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.33.attn_kv_a_norm.weight
    F32
    [512]
  • blk.33.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.33.attn_norm.weight
    F32
    [7168]
  • blk.33.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.33.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.33.attn_q_a_norm.weight
    F32
    [1536]
  • blk.33.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.33.exp_probs_b.bias
    F32
    [256]
  • blk.33.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.33.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.33.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.33.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.33.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.33.ffn_norm.weight
    F32
    [7168]
  • blk.33.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.33.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.34
  • blk.34.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.34.attn_kv_a_norm.weight
    F32
    [512]
  • blk.34.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.34.attn_norm.weight
    F32
    [7168]
  • blk.34.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.34.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.34.attn_q_a_norm.weight
    F32
    [1536]
  • blk.34.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.34.exp_probs_b.bias
    F32
    [256]
  • blk.34.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.34.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.34.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.34.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.34.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.34.ffn_norm.weight
    F32
    [7168]
  • blk.34.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.34.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.35
  • blk.35.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.35.attn_kv_a_norm.weight
    F32
    [512]
  • blk.35.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.35.attn_norm.weight
    F32
    [7168]
  • blk.35.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.35.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.35.attn_q_a_norm.weight
    F32
    [1536]
  • blk.35.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.35.exp_probs_b.bias
    F32
    [256]
  • blk.35.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.35.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.35.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.35.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.35.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.35.ffn_norm.weight
    F32
    [7168]
  • blk.35.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.35.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.36
  • blk.36.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.36.attn_kv_a_norm.weight
    F32
    [512]
  • blk.36.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.36.attn_norm.weight
    F32
    [7168]
  • blk.36.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.36.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.36.attn_q_a_norm.weight
    F32
    [1536]
  • blk.36.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.36.exp_probs_b.bias
    F32
    [256]
  • blk.36.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.36.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.36.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.36.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.36.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.36.ffn_norm.weight
    F32
    [7168]
  • blk.36.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.36.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.37
  • blk.37.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.37.attn_kv_a_norm.weight
    F32
    [512]
  • blk.37.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.37.attn_norm.weight
    F32
    [7168]
  • blk.37.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.37.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.37.attn_q_a_norm.weight
    F32
    [1536]
  • blk.37.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.37.exp_probs_b.bias
    F32
    [256]
  • blk.37.ffn_down_exps.weight
    Q2_K
    [2048, 7168, 256]
  • blk.37.ffn_down_shexp.weight
    Q6_K
    [2048, 7168]
  • blk.37.ffn_gate_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.37.ffn_gate_inp.weight
    F32
    [7168, 256]
  • blk.37.ffn_gate_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.37.ffn_norm.weight
    F32
    [7168]
  • blk.37.ffn_up_exps.weight
    Q2_K
    [7168, 2048, 256]
  • blk.37.ffn_up_shexp.weight
    Q4_K
    [7168, 2048]
  • blk.38
  • blk.38.attn_kv_a_mqa.weight
    Q6_K
    [7168, 576]
  • blk.38.attn_kv_a_norm.weight
    F32
    [512]
  • blk.38.attn_kv_b.weight
    Q6_K
    [512, 32768]
  • blk.38.attn_norm.weight
    F32
    [7168]
  • blk.38.attn_output.weight
    Q4_K
    [16384, 7168]
  • blk.38.attn_q_a.weight
    Q4_K
    [7168, 1536]
  • blk.38.attn_q_a_norm.weight
    F32
    [1536]
  • blk.38.attn_q_b.weight
    Q4_K
    [1536, 24576]
  • blk.38.exp_probs_b.bias
    F32
    [256]