Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
shivamb25
/
calme-2.1-qwen2-72b
:latest
27
Downloads
Updated
1 year ago
Cancel
calme-2.1-qwen2-72b:latest
...
/
model
4cd993c51a6b · 4.0GB
Metadata
split.count
8
8
split.no
1
1
split.tensors.count
963
963
Tensor
Name
Type
Shape
blk.9
blk.9.attn_norm.weight
F32
F32
[8192]
blk.9.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.9.ffn_norm.weight
F32
F32
[8192]
blk.9.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.10
blk.10.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.11
blk.11.attn_k.bias
F32
F32
[1024]
blk.11.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.11.attn_norm.weight
F32
F32
[8192]
blk.11.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.11.attn_q.bias
F32
F32
[8192]
blk.11.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.11.attn_v.bias
F32
F32
[1024]
blk.11.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.11.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.11.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.11.ffn_norm.weight
F32
F32
[8192]
blk.11.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.12
blk.12.attn_k.bias
F32
F32
[1024]
blk.12.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.12.attn_norm.weight
F32
F32
[8192]
blk.12.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.12.attn_q.bias
F32
F32
[8192]
blk.12.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.12.attn_v.bias
F32
F32
[1024]
blk.12.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.12.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.12.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.12.ffn_norm.weight
F32
F32
[8192]
blk.12.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.13
blk.13.attn_k.bias
F32
F32
[1024]
blk.13.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.13.attn_norm.weight
F32
F32
[8192]
blk.13.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.13.attn_q.bias
F32
F32
[8192]
blk.13.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.13.attn_v.bias
F32
F32
[1024]
blk.13.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.13.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.13.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.13.ffn_norm.weight
F32
F32
[8192]
blk.13.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.14
blk.14.attn_k.bias
F32
F32
[1024]
blk.14.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.14.attn_norm.weight
F32
F32
[8192]
blk.14.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.14.attn_q.bias
F32
F32
[8192]
blk.14.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.14.attn_v.bias
F32
F32
[1024]
blk.14.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.14.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.14.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.14.ffn_norm.weight
F32
F32
[8192]
blk.14.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.15
blk.15.attn_k.bias
F32
F32
[1024]
blk.15.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.15.attn_norm.weight
F32
F32
[8192]
blk.15.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.15.attn_q.bias
F32
F32
[8192]
blk.15.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.15.attn_v.bias
F32
F32
[1024]
blk.15.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.15.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.15.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.15.ffn_norm.weight
F32
F32
[8192]
blk.15.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.16
blk.16.attn_k.bias
F32
F32
[1024]
blk.16.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.16.attn_norm.weight
F32
F32
[8192]
blk.16.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.16.attn_q.bias
F32
F32
[8192]
blk.16.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.16.attn_v.bias
F32
F32
[1024]
blk.16.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.16.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.16.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.16.ffn_norm.weight
F32
F32
[8192]
blk.16.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.17
blk.17.attn_k.bias
F32
F32
[1024]
blk.17.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.17.attn_norm.weight
F32
F32
[8192]
blk.17.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.17.attn_q.bias
F32
F32
[8192]
blk.17.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.17.attn_v.bias
F32
F32
[1024]
blk.17.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.17.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.17.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.17.ffn_norm.weight
F32
F32
[8192]
blk.17.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.18
blk.18.attn_k.bias
F32
F32
[1024]
blk.18.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.18.attn_norm.weight
F32
F32
[8192]
blk.18.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.18.attn_q.bias
F32
F32
[8192]
blk.18.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.18.attn_v.bias
F32
F32
[1024]
blk.18.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.18.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.18.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.18.ffn_norm.weight
F32
F32
[8192]
blk.18.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.19
blk.19.attn_k.bias
F32
F32
[1024]
blk.19.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.19.attn_norm.weight
F32
F32
[8192]
blk.19.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.19.attn_q.bias
F32
F32
[8192]
blk.19.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.19.attn_v.bias
F32
F32
[1024]
blk.19.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.19.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.19.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.19.ffn_norm.weight
F32
F32
[8192]
blk.19.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.20
blk.20.attn_k.bias
F32
F32
[1024]
blk.20.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.20.attn_norm.weight
F32
F32
[8192]
blk.20.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.20.attn_q.bias
F32
F32
[8192]
blk.20.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.20.attn_v.bias
F32
F32
[1024]
blk.20.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.20.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.20.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.20.ffn_norm.weight
F32
F32
[8192]
blk.20.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.21
blk.21.attn_norm.weight
F32
F32
[8192]
blk.21.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.21.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]