Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
shivamb25
/
calme-2.1-qwen2-72b
:latest
27
Downloads
Updated
1 year ago
Cancel
calme-2.1-qwen2-72b:latest
...
/
model
9bf4101ef7ca · 3.8GB
Metadata
split.count
8
8
split.no
5
5
split.tensors.count
963
963
Tensor
Name
Type
Shape
blk.53
blk.53.attn_norm.weight
F32
F32
[8192]
blk.53.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.53.attn_q.bias
F32
F32
[8192]
blk.53.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.53.attn_v.bias
F32
F32
[1024]
blk.53.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.53.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.53.ffn_norm.weight
F32
F32
[8192]
blk.53.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.54
blk.54.attn_k.bias
F32
F32
[1024]
blk.54.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.54.attn_norm.weight
F32
F32
[8192]
blk.54.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.54.attn_q.bias
F32
F32
[8192]
blk.54.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.54.attn_v.bias
F32
F32
[1024]
blk.54.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.54.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.54.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.54.ffn_norm.weight
F32
F32
[8192]
blk.54.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.55
blk.55.attn_k.bias
F32
F32
[1024]
blk.55.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.55.attn_norm.weight
F32
F32
[8192]
blk.55.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.55.attn_q.bias
F32
F32
[8192]
blk.55.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.55.attn_v.bias
F32
F32
[1024]
blk.55.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.55.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.55.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.55.ffn_norm.weight
F32
F32
[8192]
blk.55.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.56
blk.56.attn_k.bias
F32
F32
[1024]
blk.56.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.56.attn_norm.weight
F32
F32
[8192]
blk.56.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.56.attn_q.bias
F32
F32
[8192]
blk.56.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.56.attn_v.bias
F32
F32
[1024]
blk.56.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.56.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.56.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.56.ffn_norm.weight
F32
F32
[8192]
blk.56.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.57
blk.57.attn_k.bias
F32
F32
[1024]
blk.57.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.57.attn_norm.weight
F32
F32
[8192]
blk.57.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.57.attn_q.bias
F32
F32
[8192]
blk.57.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.57.attn_v.bias
F32
F32
[1024]
blk.57.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.57.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.57.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.57.ffn_norm.weight
F32
F32
[8192]
blk.57.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.58
blk.58.attn_k.bias
F32
F32
[1024]
blk.58.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.58.attn_norm.weight
F32
F32
[8192]
blk.58.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.58.attn_q.bias
F32
F32
[8192]
blk.58.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.58.attn_v.bias
F32
F32
[1024]
blk.58.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.58.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.58.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.58.ffn_norm.weight
F32
F32
[8192]
blk.58.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.59
blk.59.attn_k.bias
F32
F32
[1024]
blk.59.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.59.attn_norm.weight
F32
F32
[8192]
blk.59.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.59.attn_q.bias
F32
F32
[8192]
blk.59.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.59.attn_v.bias
F32
F32
[1024]
blk.59.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.59.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.59.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.59.ffn_norm.weight
F32
F32
[8192]
blk.59.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.60
blk.60.attn_k.bias
F32
F32
[1024]
blk.60.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.60.attn_norm.weight
F32
F32
[8192]
blk.60.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.60.attn_q.bias
F32
F32
[8192]
blk.60.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.60.attn_v.bias
F32
F32
[1024]
blk.60.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.60.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.60.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.60.ffn_norm.weight
F32
F32
[8192]
blk.60.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.61
blk.61.attn_k.bias
F32
F32
[1024]
blk.61.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.61.attn_norm.weight
F32
F32
[8192]
blk.61.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.61.attn_q.bias
F32
F32
[8192]
blk.61.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.61.attn_v.bias
F32
F32
[1024]
blk.61.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.61.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.61.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.61.ffn_norm.weight
F32
F32
[8192]
blk.61.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.62
blk.62.attn_k.bias
F32
F32
[1024]
blk.62.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.62.attn_norm.weight
F32
F32
[8192]
blk.62.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.62.attn_q.bias
F32
F32
[8192]
blk.62.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.62.attn_v.bias
F32
F32
[1024]
blk.62.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.62.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.62.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.62.ffn_norm.weight
F32
F32
[8192]
blk.62.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.63
blk.63.attn_k.bias
F32
F32
[1024]
blk.63.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.63.attn_norm.weight
F32
F32
[8192]
blk.63.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.63.attn_q.bias
F32
F32
[8192]
blk.63.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.63.attn_v.bias
F32
F32
[1024]
blk.63.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.63.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.63.ffn_norm.weight
F32
F32
[8192]
blk.63.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]