Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
shivamb25
/
calme-2.1-qwen2-72b
:latest
27
Downloads
Updated
1 year ago
Cancel
calme-2.1-qwen2-72b:latest
...
/
model
f73f96f3b6d1 · 3.9GB
Metadata
split.count
8
8
split.no
6
6
split.tensors.count
963
963
Tensor
Name
Type
Shape
blk.63
blk.63.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.64
blk.64.attn_k.bias
F32
F32
[1024]
blk.64.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.64.attn_norm.weight
F32
F32
[8192]
blk.64.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.64.attn_q.bias
F32
F32
[8192]
blk.64.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.64.attn_v.bias
F32
F32
[1024]
blk.64.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.64.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.64.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.64.ffn_norm.weight
F32
F32
[8192]
blk.64.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.65
blk.65.attn_k.bias
F32
F32
[1024]
blk.65.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.65.attn_norm.weight
F32
F32
[8192]
blk.65.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.65.attn_q.bias
F32
F32
[8192]
blk.65.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.65.attn_v.bias
F32
F32
[1024]
blk.65.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.65.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.65.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.65.ffn_norm.weight
F32
F32
[8192]
blk.65.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.66
blk.66.attn_k.bias
F32
F32
[1024]
blk.66.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.66.attn_norm.weight
F32
F32
[8192]
blk.66.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.66.attn_q.bias
F32
F32
[8192]
blk.66.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.66.attn_v.bias
F32
F32
[1024]
blk.66.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.66.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.66.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.66.ffn_norm.weight
F32
F32
[8192]
blk.66.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.67
blk.67.attn_k.bias
F32
F32
[1024]
blk.67.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.67.attn_norm.weight
F32
F32
[8192]
blk.67.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.67.attn_q.bias
F32
F32
[8192]
blk.67.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.67.attn_v.bias
F32
F32
[1024]
blk.67.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.67.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.67.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.67.ffn_norm.weight
F32
F32
[8192]
blk.67.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.68
blk.68.attn_k.bias
F32
F32
[1024]
blk.68.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.68.attn_norm.weight
F32
F32
[8192]
blk.68.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.68.attn_q.bias
F32
F32
[8192]
blk.68.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.68.attn_v.bias
F32
F32
[1024]
blk.68.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.68.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.68.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.68.ffn_norm.weight
F32
F32
[8192]
blk.68.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.69
blk.69.attn_k.bias
F32
F32
[1024]
blk.69.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.69.attn_norm.weight
F32
F32
[8192]
blk.69.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.69.attn_q.bias
F32
F32
[8192]
blk.69.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.69.attn_v.bias
F32
F32
[1024]
blk.69.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.69.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.69.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.69.ffn_norm.weight
F32
F32
[8192]
blk.69.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.70
blk.70.attn_k.bias
F32
F32
[1024]
blk.70.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.70.attn_norm.weight
F32
F32
[8192]
blk.70.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.70.attn_q.bias
F32
F32
[8192]
blk.70.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.70.attn_v.bias
F32
F32
[1024]
blk.70.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.70.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.70.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.70.ffn_norm.weight
F32
F32
[8192]
blk.70.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.71
blk.71.attn_k.bias
F32
F32
[1024]
blk.71.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.71.attn_norm.weight
F32
F32
[8192]
blk.71.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.71.attn_q.bias
F32
F32
[8192]
blk.71.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.71.attn_v.bias
F32
F32
[1024]
blk.71.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.71.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.71.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.71.ffn_norm.weight
F32
F32
[8192]
blk.71.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.72
blk.72.attn_k.bias
F32
F32
[1024]
blk.72.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.72.attn_norm.weight
F32
F32
[8192]
blk.72.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.72.attn_q.bias
F32
F32
[8192]
blk.72.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.72.attn_v.bias
F32
F32
[1024]
blk.72.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.72.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.72.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.72.ffn_norm.weight
F32
F32
[8192]
blk.72.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.73
blk.73.attn_k.bias
F32
F32
[1024]
blk.73.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.73.attn_norm.weight
F32
F32
[8192]
blk.73.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.73.attn_q.bias
F32
F32
[8192]
blk.73.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.73.attn_v.bias
F32
F32
[1024]
blk.73.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.73.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.73.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.73.ffn_norm.weight
F32
F32
[8192]
blk.73.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.74
blk.74.attn_k.bias
F32
F32
[1024]
blk.74.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.74.attn_norm.weight
F32
F32
[8192]
blk.74.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.74.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.74.ffn_norm.weight
F32
F32
[8192]
blk.74.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]