Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
shivamb25
/
calme-2.1-qwen2-72b
:latest
27
Downloads
Updated
1 year ago
Cancel
calme-2.1-qwen2-72b:latest
...
/
model
6fef0021bcb4 · 2.8GB
Metadata
split.count
8
8
split.no
7
7
split.tensors.count
963
963
Tensor
Name
Type
Shape
blk.74
blk.74.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.74.attn_q.bias
F32
F32
[8192]
blk.74.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.74.attn_v.bias
F32
F32
[1024]
blk.74.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.75
blk.75.attn_k.bias
F32
F32
[1024]
blk.75.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.75.attn_norm.weight
F32
F32
[8192]
blk.75.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.75.attn_q.bias
F32
F32
[8192]
blk.75.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.75.attn_v.bias
F32
F32
[1024]
blk.75.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.75.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.75.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.75.ffn_norm.weight
F32
F32
[8192]
blk.75.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.76
blk.76.attn_k.bias
F32
F32
[1024]
blk.76.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.76.attn_norm.weight
F32
F32
[8192]
blk.76.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.76.attn_q.bias
F32
F32
[8192]
blk.76.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.76.attn_v.bias
F32
F32
[1024]
blk.76.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.76.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.76.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.76.ffn_norm.weight
F32
F32
[8192]
blk.76.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.77
blk.77.attn_k.bias
F32
F32
[1024]
blk.77.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.77.attn_norm.weight
F32
F32
[8192]
blk.77.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.77.attn_q.bias
F32
F32
[8192]
blk.77.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.77.attn_v.bias
F32
F32
[1024]
blk.77.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.77.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.77.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.77.ffn_norm.weight
F32
F32
[8192]
blk.77.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.78
blk.78.attn_k.bias
F32
F32
[1024]
blk.78.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.78.attn_norm.weight
F32
F32
[8192]
blk.78.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.78.attn_q.bias
F32
F32
[8192]
blk.78.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.78.attn_v.bias
F32
F32
[1024]
blk.78.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.78.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.78.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.78.ffn_norm.weight
F32
F32
[8192]
blk.78.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.79
blk.79.attn_k.bias
F32
F32
[1024]
blk.79.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.79.attn_norm.weight
F32
F32
[8192]
blk.79.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.79.attn_q.bias
F32
F32
[8192]
blk.79.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.79.attn_v.bias
F32
F32
[1024]
blk.79.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.79.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.79.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.79.ffn_norm.weight
F32
F32
[8192]
blk.79.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
output.weight
Q6_K
Q6_K
[8192, 151646]
output_norm.weight
F32
F32
[8192]