27 1 year ago

6fef0021bcb4 · 2.8GB
    Metadata
  • split.count
    8
  • split.no
    7
  • split.tensors.count
    963
  • Tensor
    blk.74
  • blk.74.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.74.attn_q.bias
    F32
    [8192]
  • blk.74.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.74.attn_v.bias
    F32
    [1024]
  • blk.74.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.75
  • blk.75.attn_k.bias
    F32
    [1024]
  • blk.75.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.75.attn_norm.weight
    F32
    [8192]
  • blk.75.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.75.attn_q.bias
    F32
    [8192]
  • blk.75.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.75.attn_v.bias
    F32
    [1024]
  • blk.75.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.75.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.75.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.75.ffn_norm.weight
    F32
    [8192]
  • blk.75.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.76
  • blk.76.attn_k.bias
    F32
    [1024]
  • blk.76.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.76.attn_norm.weight
    F32
    [8192]
  • blk.76.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.76.attn_q.bias
    F32
    [8192]
  • blk.76.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.76.attn_v.bias
    F32
    [1024]
  • blk.76.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.76.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.76.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.76.ffn_norm.weight
    F32
    [8192]
  • blk.76.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.77
  • blk.77.attn_k.bias
    F32
    [1024]
  • blk.77.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.77.attn_norm.weight
    F32
    [8192]
  • blk.77.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.77.attn_q.bias
    F32
    [8192]
  • blk.77.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.77.attn_v.bias
    F32
    [1024]
  • blk.77.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.77.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.77.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.77.ffn_norm.weight
    F32
    [8192]
  • blk.77.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.78
  • blk.78.attn_k.bias
    F32
    [1024]
  • blk.78.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.78.attn_norm.weight
    F32
    [8192]
  • blk.78.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.78.attn_q.bias
    F32
    [8192]
  • blk.78.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.78.attn_v.bias
    F32
    [1024]
  • blk.78.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.78.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.78.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.78.ffn_norm.weight
    F32
    [8192]
  • blk.78.ffn_up.weight
    Q2_K
    [8192, 29568]
  • blk.79
  • blk.79.attn_k.bias
    F32
    [1024]
  • blk.79.attn_k.weight
    Q2_K
    [8192, 1024]
  • blk.79.attn_norm.weight
    F32
    [8192]
  • blk.79.attn_output.weight
    Q3_K
    [8192, 8192]
  • blk.79.attn_q.bias
    F32
    [8192]
  • blk.79.attn_q.weight
    Q2_K
    [8192, 8192]
  • blk.79.attn_v.bias
    F32
    [1024]
  • blk.79.attn_v.weight
    Q5_K
    [8192, 1024]
  • blk.79.ffn_down.weight
    IQ4_NL
    [29568, 8192]
  • blk.79.ffn_gate.weight
    Q2_K
    [8192, 29568]
  • blk.79.ffn_norm.weight
    F32
    [8192]
  • blk.79.ffn_up.weight
    Q2_K
    [8192, 29568]
  • output.weight
    Q6_K
    [8192, 151646]
  • output_norm.weight
    F32
    [8192]