Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
shivamb25
/
calme-2.1-qwen2-72b
:latest
27
Downloads
Updated
1 year ago
Cancel
calme-2.1-qwen2-72b:latest
...
/
model
9ae63a84c82c · 3.7GB
Metadata
split.count
8
8
split.no
2
2
split.tensors.count
963
963
Tensor
Name
Type
Shape
blk.21
blk.21.attn_k.bias
F32
F32
[1024]
blk.21.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.21.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.21.attn_q.bias
F32
F32
[8192]
blk.21.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.21.attn_v.bias
F32
F32
[1024]
blk.21.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.21.ffn_norm.weight
F32
F32
[8192]
blk.21.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.22
blk.22.attn_k.bias
F32
F32
[1024]
blk.22.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.22.attn_norm.weight
F32
F32
[8192]
blk.22.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.22.attn_q.bias
F32
F32
[8192]
blk.22.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.22.attn_v.bias
F32
F32
[1024]
blk.22.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.22.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.22.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.22.ffn_norm.weight
F32
F32
[8192]
blk.22.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.23
blk.23.attn_k.bias
F32
F32
[1024]
blk.23.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.23.attn_norm.weight
F32
F32
[8192]
blk.23.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.23.attn_q.bias
F32
F32
[8192]
blk.23.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.23.attn_v.bias
F32
F32
[1024]
blk.23.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.23.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.23.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.23.ffn_norm.weight
F32
F32
[8192]
blk.23.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.24
blk.24.attn_k.bias
F32
F32
[1024]
blk.24.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.24.attn_norm.weight
F32
F32
[8192]
blk.24.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.24.attn_q.bias
F32
F32
[8192]
blk.24.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.24.attn_v.bias
F32
F32
[1024]
blk.24.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.24.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.24.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.24.ffn_norm.weight
F32
F32
[8192]
blk.24.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.25
blk.25.attn_k.bias
F32
F32
[1024]
blk.25.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.25.attn_norm.weight
F32
F32
[8192]
blk.25.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.25.attn_q.bias
F32
F32
[8192]
blk.25.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.25.attn_v.bias
F32
F32
[1024]
blk.25.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.25.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.25.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.25.ffn_norm.weight
F32
F32
[8192]
blk.25.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.26
blk.26.attn_k.bias
F32
F32
[1024]
blk.26.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.26.attn_norm.weight
F32
F32
[8192]
blk.26.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.26.attn_q.bias
F32
F32
[8192]
blk.26.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.26.attn_v.bias
F32
F32
[1024]
blk.26.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.26.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.26.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.26.ffn_norm.weight
F32
F32
[8192]
blk.26.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.27
blk.27.attn_k.bias
F32
F32
[1024]
blk.27.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.27.attn_norm.weight
F32
F32
[8192]
blk.27.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.27.attn_q.bias
F32
F32
[8192]
blk.27.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.27.attn_v.bias
F32
F32
[1024]
blk.27.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.27.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.27.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.27.ffn_norm.weight
F32
F32
[8192]
blk.27.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.28
blk.28.attn_k.bias
F32
F32
[1024]
blk.28.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.28.attn_norm.weight
F32
F32
[8192]
blk.28.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.28.attn_q.bias
F32
F32
[8192]
blk.28.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.28.attn_v.bias
F32
F32
[1024]
blk.28.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.28.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.28.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.28.ffn_norm.weight
F32
F32
[8192]
blk.28.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.29
blk.29.attn_k.bias
F32
F32
[1024]
blk.29.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.29.attn_norm.weight
F32
F32
[8192]
blk.29.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.29.attn_q.bias
F32
F32
[8192]
blk.29.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.29.attn_v.bias
F32
F32
[1024]
blk.29.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.29.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.29.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.29.ffn_norm.weight
F32
F32
[8192]
blk.29.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.30
blk.30.attn_k.bias
F32
F32
[1024]
blk.30.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.30.attn_norm.weight
F32
F32
[8192]
blk.30.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.30.attn_q.bias
F32
F32
[8192]
blk.30.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.30.attn_v.bias
F32
F32
[1024]
blk.30.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.30.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.30.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.30.ffn_norm.weight
F32
F32
[8192]
blk.30.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.31
blk.31.attn_k.bias
F32
F32
[1024]
blk.31.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.31.attn_norm.weight
F32
F32
[8192]
blk.31.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.31.attn_q.bias
F32
F32
[8192]
blk.31.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.31.attn_v.bias
F32
F32
[1024]
blk.31.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.31.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.31.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.31.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]