Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
shivamb25
/
calme-2.1-qwen2-72b
:latest
27
Downloads
Updated
1 year ago
Cancel
calme-2.1-qwen2-72b:latest
...
/
model
7f0f12151314 · 3.7GB
Metadata
split.count
8
8
split.no
3
3
split.tensors.count
963
963
Tensor
Name
Type
Shape
blk.31
blk.31.ffn_norm.weight
F32
F32
[8192]
blk.32
blk.32.attn_k.bias
F32
F32
[1024]
blk.32.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.32.attn_norm.weight
F32
F32
[8192]
blk.32.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.32.attn_q.bias
F32
F32
[8192]
blk.32.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.32.attn_v.bias
F32
F32
[1024]
blk.32.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.32.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.32.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.32.ffn_norm.weight
F32
F32
[8192]
blk.32.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.33
blk.33.attn_k.bias
F32
F32
[1024]
blk.33.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.33.attn_norm.weight
F32
F32
[8192]
blk.33.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.33.attn_q.bias
F32
F32
[8192]
blk.33.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.33.attn_v.bias
F32
F32
[1024]
blk.33.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.33.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.33.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.33.ffn_norm.weight
F32
F32
[8192]
blk.33.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.34
blk.34.attn_k.bias
F32
F32
[1024]
blk.34.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.34.attn_norm.weight
F32
F32
[8192]
blk.34.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.34.attn_q.bias
F32
F32
[8192]
blk.34.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.34.attn_v.bias
F32
F32
[1024]
blk.34.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.34.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.34.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.34.ffn_norm.weight
F32
F32
[8192]
blk.34.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.35
blk.35.attn_k.bias
F32
F32
[1024]
blk.35.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.35.attn_norm.weight
F32
F32
[8192]
blk.35.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.35.attn_q.bias
F32
F32
[8192]
blk.35.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.35.attn_v.bias
F32
F32
[1024]
blk.35.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.35.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.35.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.35.ffn_norm.weight
F32
F32
[8192]
blk.35.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.36
blk.36.attn_k.bias
F32
F32
[1024]
blk.36.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.36.attn_norm.weight
F32
F32
[8192]
blk.36.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.36.attn_q.bias
F32
F32
[8192]
blk.36.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.36.attn_v.bias
F32
F32
[1024]
blk.36.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.36.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.36.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.36.ffn_norm.weight
F32
F32
[8192]
blk.36.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.37
blk.37.attn_k.bias
F32
F32
[1024]
blk.37.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.37.attn_norm.weight
F32
F32
[8192]
blk.37.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.37.attn_q.bias
F32
F32
[8192]
blk.37.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.37.attn_v.bias
F32
F32
[1024]
blk.37.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.37.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.37.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.37.ffn_norm.weight
F32
F32
[8192]
blk.37.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.38
blk.38.attn_k.bias
F32
F32
[1024]
blk.38.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.38.attn_norm.weight
F32
F32
[8192]
blk.38.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.38.attn_q.bias
F32
F32
[8192]
blk.38.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.38.attn_v.bias
F32
F32
[1024]
blk.38.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.38.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.38.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.38.ffn_norm.weight
F32
F32
[8192]
blk.38.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.39
blk.39.attn_k.bias
F32
F32
[1024]
blk.39.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.39.attn_norm.weight
F32
F32
[8192]
blk.39.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.39.attn_q.bias
F32
F32
[8192]
blk.39.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.39.attn_v.bias
F32
F32
[1024]
blk.39.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.39.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.39.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.39.ffn_norm.weight
F32
F32
[8192]
blk.39.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.40
blk.40.attn_k.bias
F32
F32
[1024]
blk.40.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.40.attn_norm.weight
F32
F32
[8192]
blk.40.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.40.attn_q.bias
F32
F32
[8192]
blk.40.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.40.attn_v.bias
F32
F32
[1024]
blk.40.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.40.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.40.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.40.ffn_norm.weight
F32
F32
[8192]
blk.40.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.41
blk.41.attn_k.bias
F32
F32
[1024]
blk.41.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.41.attn_norm.weight
F32
F32
[8192]
blk.41.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.41.attn_q.bias
F32
F32
[8192]
blk.41.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.41.attn_v.bias
F32
F32
[1024]
blk.41.attn_v.weight
Q5_K
Q5_K
[8192, 1024]
blk.41.ffn_down.weight
IQ4_NL
IQ4_NL
[29568, 8192]
blk.41.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]
blk.41.ffn_norm.weight
F32
F32
[8192]
blk.41.ffn_up.weight
Q2_K
Q2_K
[8192, 29568]
blk.42
blk.42.attn_k.bias
F32
F32
[1024]
blk.42.attn_k.weight
Q2_K
Q2_K
[8192, 1024]
blk.42.attn_output.weight
Q3_K
Q3_K
[8192, 8192]
blk.42.attn_q.bias
F32
F32
[8192]
blk.42.attn_q.weight
Q2_K
Q2_K
[8192, 8192]
blk.42.attn_v.bias
F32
F32
[1024]
blk.42.ffn_gate.weight
Q2_K
Q2_K
[8192, 29568]