27 1 year ago

9ae63a84c82c · 3.7GB
    Metadata
  • split.count
    8
  • split.no
    2
  • split.tensors.count
    963
  • Tensor
    blk.21
  • blk.21.attn_k.bias
    F32
    [1024]
  • blk.21.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.21.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.21.attn_q.bias
    F32
    [8192]
  • blk.21.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.21.attn_v.bias
    F32
    [1024]
  • blk.21.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.21.ffn_norm.weight
    F32
    [8192]
  • blk.21.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.22
  • blk.22.attn_k.bias
    F32
    [1024]
  • blk.22.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.22.attn_norm.weight
    F32
    [8192]
  • blk.22.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.22.attn_q.bias
    F32
    [8192]
  • blk.22.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.22.attn_v.bias
    F32
    [1024]
  • blk.22.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.22.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.22.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.22.ffn_norm.weight
    F32
    [8192]
  • blk.22.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.23
  • blk.23.attn_k.bias
    F32
    [1024]
  • blk.23.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.23.attn_norm.weight
    F32
    [8192]
  • blk.23.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.23.attn_q.bias
    F32
    [8192]
  • blk.23.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.23.attn_v.bias
    F32
    [1024]
  • blk.23.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.23.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.23.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.23.ffn_norm.weight
    F32
    [8192]
  • blk.23.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.24
  • blk.24.attn_k.bias
    F32
    [1024]
  • blk.24.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.24.attn_norm.weight
    F32
    [8192]
  • blk.24.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.24.attn_q.bias
    F32
    [8192]
  • blk.24.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.24.attn_v.bias
    F32
    [1024]
  • blk.24.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.24.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.24.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.24.ffn_norm.weight
    F32
    [8192]
  • blk.24.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.25
  • blk.25.attn_k.bias
    F32
    [1024]
  • blk.25.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.25.attn_norm.weight
    F32
    [8192]
  • blk.25.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.25.attn_q.bias
    F32
    [8192]
  • blk.25.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.25.attn_v.bias
    F32
    [1024]
  • blk.25.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.25.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.25.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.25.ffn_norm.weight
    F32
    [8192]
  • blk.25.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.26
  • blk.26.attn_k.bias
    F32
    [1024]
  • blk.26.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.26.attn_norm.weight
    F32
    [8192]
  • blk.26.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.26.attn_q.bias
    F32
    [8192]
  • blk.26.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.26.attn_v.bias
    F32
    [1024]
  • blk.26.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.26.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.26.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.26.ffn_norm.weight
    F32
    [8192]
  • blk.26.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.27
  • blk.27.attn_k.bias
    F32
    [1024]
  • blk.27.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.27.attn_norm.weight
    F32
    [8192]
  • blk.27.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.27.attn_q.bias
    F32
    [8192]
  • blk.27.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.27.attn_v.bias
    F32
    [1024]
  • blk.27.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.27.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.27.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.27.ffn_norm.weight
    F32
    [8192]
  • blk.27.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.28
  • blk.28.attn_k.bias
    F32
    [1024]
  • blk.28.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.28.attn_norm.weight
    F32
    [8192]
  • blk.28.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.28.attn_q.bias
    F32
    [8192]
  • blk.28.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.28.attn_v.bias
    F32
    [1024]
  • blk.28.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.28.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.28.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.28.ffn_norm.weight
    F32
    [8192]
  • blk.28.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.29
  • blk.29.attn_k.bias
    F32
    [1024]
  • blk.29.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.29.attn_norm.weight
    F32
    [8192]
  • blk.29.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.29.attn_q.bias
    F32
    [8192]
  • blk.29.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.29.attn_v.bias
    F32
    [1024]
  • blk.29.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.29.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.29.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.29.ffn_norm.weight
    F32
    [8192]
  • blk.29.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.30
  • blk.30.attn_k.bias
    F32
    [1024]
  • blk.30.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.30.attn_norm.weight
    F32
    [8192]
  • blk.30.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.30.attn_q.bias
    F32
    [8192]
  • blk.30.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.30.attn_v.bias
    F32
    [1024]
  • blk.30.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.30.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.30.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.30.ffn_norm.weight
    F32
    [8192]
  • blk.30.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.31
  • blk.31.attn_k.bias
    F32
    [1024]
  • blk.31.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.31.attn_norm.weight
    F32
    [8192]
  • blk.31.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.31.attn_q.bias
    F32
    [8192]
  • blk.31.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.31.attn_v.bias
    F32
    [1024]
  • blk.31.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.31.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.31.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.31.ffn_up.weight
    Q2_K
    [8192, 29568]