Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
shivamb25
/
calme-2.1-qwen2-72b
:latest
27
Downloads
Updated
1 year ago
Cancel
calme-2.1-qwen2-72b:latest
...
/
model
02a129b9b26f · 3.9GB
Metadata
split.count
8
8
split.no
4
4
split.tensors.count
963
963
Tensor
Name
Type
Shape
blk.42
blk.42.attn_norm.weight
F32
F32
[8192]
blk.42.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.42.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.42.ffn_norm.weight
F32
F32
[8192]
blk.42.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.43
blk.43.attn_k.bias
F32
F32
[1024]
blk.43.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.43.attn_norm.weight
F32
F32
[8192]
blk.43.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.43.attn_q.bias
F32
F32
[8192]
blk.43.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.43.attn_v.bias
F32
F32
[1024]
blk.43.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.43.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.43.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.43.ffn_norm.weight
F32
F32
[8192]
blk.43.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.44
blk.44.attn_k.bias
F32
F32
[1024]
blk.44.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.44.attn_norm.weight
F32
F32
[8192]
blk.44.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.44.attn_q.bias
F32
F32
[8192]
blk.44.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.44.attn_v.bias
F32
F32
[1024]
blk.44.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.44.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.44.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.44.ffn_norm.weight
F32
F32
[8192]
blk.44.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.45
blk.45.attn_k.bias
F32
F32
[1024]
blk.45.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.45.attn_norm.weight
F32
F32
[8192]
blk.45.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.45.attn_q.bias
F32
F32
[8192]
blk.45.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.45.attn_v.bias
F32
F32
[1024]
blk.45.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.45.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.45.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.45.ffn_norm.weight
F32
F32
[8192]
blk.45.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.46
blk.46.attn_k.bias
F32
F32
[1024]
blk.46.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.46.attn_norm.weight
F32
F32
[8192]
blk.46.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.46.attn_q.bias
F32
F32
[8192]
blk.46.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.46.attn_v.bias
F32
F32
[1024]
blk.46.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.46.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.46.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.46.ffn_norm.weight
F32
F32
[8192]
blk.46.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.47
blk.47.attn_k.bias
F32
F32
[1024]
blk.47.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.47.attn_norm.weight
F32
F32
[8192]
blk.47.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.47.attn_q.bias
F32
F32
[8192]
blk.47.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.47.attn_v.bias
F32
F32
[1024]
blk.47.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.47.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.47.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.47.ffn_norm.weight
F32
F32
[8192]
blk.47.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.48
blk.48.attn_k.bias
F32
F32
[1024]
blk.48.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.48.attn_norm.weight
F32
F32
[8192]
blk.48.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.48.attn_q.bias
F32
F32
[8192]
blk.48.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.48.attn_v.bias
F32
F32
[1024]
blk.48.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.48.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.48.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.48.ffn_norm.weight
F32
F32
[8192]
blk.48.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.49
blk.49.attn_k.bias
F32
F32
[1024]
blk.49.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.49.attn_norm.weight
F32
F32
[8192]
blk.49.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.49.attn_q.bias
F32
F32
[8192]
blk.49.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.49.attn_v.bias
F32
F32
[1024]
blk.49.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.49.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.49.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.49.ffn_norm.weight
F32
F32
[8192]
blk.49.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.50
blk.50.attn_k.bias
F32
F32
[1024]
blk.50.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.50.attn_norm.weight
F32
F32
[8192]
blk.50.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.50.attn_q.bias
F32
F32
[8192]
blk.50.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.50.attn_v.bias
F32
F32
[1024]
blk.50.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.50.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.50.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.50.ffn_norm.weight
F32
F32
[8192]
blk.50.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.51
blk.51.attn_k.bias
F32
F32
[1024]
blk.51.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.51.attn_norm.weight
F32
F32
[8192]
blk.51.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.51.attn_q.bias
F32
F32
[8192]
blk.51.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.51.attn_v.bias
F32
F32
[1024]
blk.51.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.51.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.51.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.51.ffn_norm.weight
F32
F32
[8192]
blk.51.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.52
blk.52.attn_k.bias
F32
F32
[1024]
blk.52.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.52.attn_norm.weight
F32
F32
[8192]
blk.52.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.52.attn_q.bias
F32
F32
[8192]
blk.52.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.52.attn_v.bias
F32
F32
[1024]
blk.52.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.52.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.52.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.52.ffn_norm.weight
F32
F32
[8192]
blk.52.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.53
blk.53.attn_k.bias
F32
F32
[1024]
blk.53.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.53.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]