Models
GitHub
Discord
Docs
Pricing
Sign in
Download
Models
Download
GitHub
Discord
Docs
Pricing
Sign in
glm-ocr
:bf16
4
Downloads
Updated
an hour ago
GLM-OCR is a multimodal OCR model for complex document understanding, built on the GLM-V encoder–decoder architecture.
GLM-OCR is a multimodal OCR model for complex document understanding, built on the GLM-V encoder–decoder architecture.
Cancel
vision
glm-ocr:bf16
...
/
model
65493e1f85b9 · 2.2GB
Metadata
general.architecture
glmocr
glmocr
general.file_type
F16
F16
glmocr.attention.head_count
16
16
glmocr.attention.head_count_kv
8
8
glmocr.attention.key_length
128
128
glmocr.attention.layer_norm_rms_epsilon
1e-05
1e-05
glmocr.attention.value_length
128
128
glmocr.block_count
16
16
glmocr.context_length
131072
131072
glmocr.embedding_length
1536
1536
glmocr.feed_forward_length
4608
4608
glmocr.image_end_token_id
59257
59257
glmocr.image_start_token_id
59256
59256
glmocr.image_token_id
59280
59280
glmocr.rope.freq_base
10000
10000
glmocr.rope.mrope_section
[16, 24, 24]
[16, 24, 24]
glmocr.rope.partial_rotary_factor
1
1
glmocr.video_end_token_id
59259
59259
glmocr.video_start_token_id
59258
59258
glmocr.video_token_id
59281
59281
glmocr.vision.attention.head_count
16
16
glmocr.vision.attention.layer_norm_rms_epsilon
1e-05
1e-05
glmocr.vision.block_count
24
24
glmocr.vision.embedding_length
1024
1024
glmocr.vision.image_mean
[0.48145467, 0.4578275, 0.40821072]
[0.48145467, 0.4578275, 0.40821072]
glmocr.vision.image_size
336
336
glmocr.vision.image_std
[0.26862955, 0.2613026, 0.2757771]
[0.26862955, 0.2613026, 0.2757771]
glmocr.vision.intermediate_size
4096
4096
glmocr.vision.max_pixels
9633792
9633792
glmocr.vision.min_pixels
12544
12544
glmocr.vision.out_hidden_size
1536
1536
glmocr.vision.patch_size
14
14
glmocr.vision.spatial_merge_size
2
2
glmocr.vision.temporal_patch_size
2
2
tokenizer.ggml.add_eos_token
false
false
tokenizer.ggml.add_padding_token
false
false
tokenizer.ggml.eos_token_id
59246
59246
tokenizer.ggml.eos_token_ids
[59246, 59253]
[59246, 59253]
tokenizer.ggml.merges
[Ġ Ġ, Ġ ĠĠĠ, ĠĠ ĠĠ, ĠĠĠ Ġ, Ġ t, ...]
[Ġ Ġ, Ġ ĠĠĠ, ĠĠ ĠĠ, ĠĠĠ Ġ, Ġ t, ...]
tokenizer.ggml.model
gpt2
gpt2
tokenizer.ggml.padding_token_id
59246
59246
tokenizer.ggml.pre
llama-bpe
llama-bpe
tokenizer.ggml.scores
[0, 1, 2, 3, 4, ...]
[0, 1, 2, 3, 4, ...]
tokenizer.ggml.token_type
[1, 1, 1, 1, 1, ...]
[1, 1, 1, 1, 1, ...]
tokenizer.ggml.tokens
[Ā, ā, Ă, ă, Ą, ...]
[Ā, ā, Ă, ă, Ą, ...]
Tensor
Name
Type
Shape
token_embd.weight
BF16
BF16
[1536, 59392]
blk.0
blk.0.attn_k.weight
BF16
BF16
[1536, 1024]
blk.0.attn_norm.weight
F32
F32
[1536]
blk.0.attn_out.weight
BF16
BF16
[2048, 1536]
blk.0.attn_q.weight
BF16
BF16
[1536, 2048]
blk.0.attn_v.weight
BF16
BF16
[1536, 1024]
blk.0.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.0.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.0.ffn_norm.weight
F32
F32
[1536]
blk.0.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.0.post_attn_norm.weight
F32
F32
[1536]
blk.0.post_ffn_norm.weight
F32
F32
[1536]
blk.1
blk.1.attn_k.weight
BF16
BF16
[1536, 1024]
blk.1.attn_norm.weight
F32
F32
[1536]
blk.1.attn_out.weight
BF16
BF16
[2048, 1536]
blk.1.attn_q.weight
BF16
BF16
[1536, 2048]
blk.1.attn_v.weight
BF16
BF16
[1536, 1024]
blk.1.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.1.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.1.ffn_norm.weight
F32
F32
[1536]
blk.1.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.1.post_attn_norm.weight
F32
F32
[1536]
blk.1.post_ffn_norm.weight
F32
F32
[1536]
blk.2
blk.2.attn_k.weight
BF16
BF16
[1536, 1024]
blk.2.attn_norm.weight
F32
F32
[1536]
blk.2.attn_out.weight
BF16
BF16
[2048, 1536]
blk.2.attn_q.weight
BF16
BF16
[1536, 2048]
blk.2.attn_v.weight
BF16
BF16
[1536, 1024]
blk.2.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.2.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.2.ffn_norm.weight
F32
F32
[1536]
blk.2.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.2.post_attn_norm.weight
F32
F32
[1536]
blk.2.post_ffn_norm.weight
F32
F32
[1536]
blk.3
blk.3.attn_k.weight
BF16
BF16
[1536, 1024]
blk.3.attn_norm.weight
F32
F32
[1536]
blk.3.attn_out.weight
BF16
BF16
[2048, 1536]
blk.3.attn_q.weight
BF16
BF16
[1536, 2048]
blk.3.attn_v.weight
BF16
BF16
[1536, 1024]
blk.3.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.3.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.3.ffn_norm.weight
F32
F32
[1536]
blk.3.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.3.post_attn_norm.weight
F32
F32
[1536]
blk.3.post_ffn_norm.weight
F32
F32
[1536]
blk.4
blk.4.attn_k.weight
BF16
BF16
[1536, 1024]
blk.4.attn_norm.weight
F32
F32
[1536]
blk.4.attn_out.weight
BF16
BF16
[2048, 1536]
blk.4.attn_q.weight
BF16
BF16
[1536, 2048]
blk.4.attn_v.weight
BF16
BF16
[1536, 1024]
blk.4.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.4.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.4.ffn_norm.weight
F32
F32
[1536]
blk.4.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.4.post_attn_norm.weight
F32
F32
[1536]
blk.4.post_ffn_norm.weight
F32
F32
[1536]
blk.5
blk.5.attn_k.weight
BF16
BF16
[1536, 1024]
blk.5.attn_norm.weight
F32
F32
[1536]
blk.5.attn_out.weight
BF16
BF16
[2048, 1536]
blk.5.attn_q.weight
BF16
BF16
[1536, 2048]
blk.5.attn_v.weight
BF16
BF16
[1536, 1024]
blk.5.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.5.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.5.ffn_norm.weight
F32
F32
[1536]
blk.5.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.5.post_attn_norm.weight
F32
F32
[1536]
blk.5.post_ffn_norm.weight
F32
F32
[1536]
blk.6
blk.6.attn_k.weight
BF16
BF16
[1536, 1024]
blk.6.attn_norm.weight
F32
F32
[1536]
blk.6.attn_out.weight
BF16
BF16
[2048, 1536]
blk.6.attn_q.weight
BF16
BF16
[1536, 2048]
blk.6.attn_v.weight
BF16
BF16
[1536, 1024]
blk.6.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.6.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.6.ffn_norm.weight
F32
F32
[1536]
blk.6.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.6.post_attn_norm.weight
F32
F32
[1536]
blk.6.post_ffn_norm.weight
F32
F32
[1536]
blk.7
blk.7.attn_k.weight
BF16
BF16
[1536, 1024]
blk.7.attn_norm.weight
F32
F32
[1536]
blk.7.attn_out.weight
BF16
BF16
[2048, 1536]
blk.7.attn_q.weight
BF16
BF16
[1536, 2048]
blk.7.attn_v.weight
BF16
BF16
[1536, 1024]
blk.7.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.7.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.7.ffn_norm.weight
F32
F32
[1536]
blk.7.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.7.post_attn_norm.weight
F32
F32
[1536]
blk.7.post_ffn_norm.weight
F32
F32
[1536]
blk.8
blk.8.attn_k.weight
BF16
BF16
[1536, 1024]
blk.8.attn_norm.weight
F32
F32
[1536]
blk.8.attn_out.weight
BF16
BF16
[2048, 1536]
blk.8.attn_q.weight
BF16
BF16
[1536, 2048]
blk.8.attn_v.weight
BF16
BF16
[1536, 1024]
blk.8.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.8.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.8.ffn_norm.weight
F32
F32
[1536]
blk.8.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.8.post_attn_norm.weight
F32
F32
[1536]
blk.8.post_ffn_norm.weight
F32
F32
[1536]
blk.9
blk.9.attn_k.weight
BF16
BF16
[1536, 1024]
blk.9.attn_norm.weight
F32
F32
[1536]
blk.9.attn_out.weight
BF16
BF16
[2048, 1536]
blk.9.attn_q.weight
BF16
BF16
[1536, 2048]
blk.9.attn_v.weight
BF16
BF16
[1536, 1024]
blk.9.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.9.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.9.ffn_norm.weight
F32
F32
[1536]
blk.9.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.9.post_attn_norm.weight
F32
F32
[1536]
blk.9.post_ffn_norm.weight
F32
F32
[1536]
blk.10
blk.10.attn_k.weight
BF16
BF16
[1536, 1024]
blk.10.attn_norm.weight
F32
F32
[1536]
blk.10.attn_out.weight
BF16
BF16
[2048, 1536]
blk.10.attn_q.weight
BF16
BF16
[1536, 2048]
blk.10.attn_v.weight
BF16
BF16
[1536, 1024]
blk.10.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.10.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.10.ffn_norm.weight
F32
F32
[1536]
blk.10.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.10.post_attn_norm.weight
F32
F32
[1536]
blk.10.post_ffn_norm.weight
F32
F32
[1536]
blk.11
blk.11.attn_k.weight
BF16
BF16
[1536, 1024]
blk.11.attn_norm.weight
F32
F32
[1536]
blk.11.attn_out.weight
BF16
BF16
[2048, 1536]
blk.11.attn_q.weight
BF16
BF16
[1536, 2048]
blk.11.attn_v.weight
BF16
BF16
[1536, 1024]
blk.11.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.11.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.11.ffn_norm.weight
F32
F32
[1536]
blk.11.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.11.post_attn_norm.weight
F32
F32
[1536]
blk.11.post_ffn_norm.weight
F32
F32
[1536]
blk.12
blk.12.attn_k.weight
BF16
BF16
[1536, 1024]
blk.12.attn_norm.weight
F32
F32
[1536]
blk.12.attn_out.weight
BF16
BF16
[2048, 1536]
blk.12.attn_q.weight
BF16
BF16
[1536, 2048]
blk.12.attn_v.weight
BF16
BF16
[1536, 1024]
blk.12.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.12.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.12.ffn_norm.weight
F32
F32
[1536]
blk.12.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.12.post_attn_norm.weight
F32
F32
[1536]
blk.12.post_ffn_norm.weight
F32
F32
[1536]
blk.13
blk.13.attn_k.weight
BF16
BF16
[1536, 1024]
blk.13.attn_norm.weight
F32
F32
[1536]
blk.13.attn_out.weight
BF16
BF16
[2048, 1536]
blk.13.attn_q.weight
BF16
BF16
[1536, 2048]
blk.13.attn_v.weight
BF16
BF16
[1536, 1024]
blk.13.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.13.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.13.ffn_norm.weight
F32
F32
[1536]
blk.13.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.13.post_attn_norm.weight
F32
F32
[1536]
blk.13.post_ffn_norm.weight
F32
F32
[1536]
blk.14
blk.14.attn_k.weight
BF16
BF16
[1536, 1024]
blk.14.attn_norm.weight
F32
F32
[1536]
blk.14.attn_out.weight
BF16
BF16
[2048, 1536]
blk.14.attn_q.weight
BF16
BF16
[1536, 2048]
blk.14.attn_v.weight
BF16
BF16
[1536, 1024]
blk.14.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.14.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.14.ffn_norm.weight
F32
F32
[1536]
blk.14.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.14.post_attn_norm.weight
F32
F32
[1536]
blk.14.post_ffn_norm.weight
F32
F32
[1536]
blk.15
blk.15.attn_k.weight
BF16
BF16
[1536, 1024]
blk.15.attn_norm.weight
F32
F32
[1536]
blk.15.attn_out.weight
BF16
BF16
[2048, 1536]
blk.15.attn_q.weight
BF16
BF16
[1536, 2048]
blk.15.attn_v.weight
BF16
BF16
[1536, 1024]
blk.15.ffn_down.weight
BF16
BF16
[4608, 1536]
blk.15.ffn_gate.weight
BF16
BF16
[1536, 4608]
blk.15.ffn_norm.weight
F32
F32
[1536]
blk.15.ffn_up.weight
BF16
BF16
[1536, 4608]
blk.15.post_attn_norm.weight
F32
F32
[1536]
blk.15.post_ffn_norm.weight
F32
F32
[1536]
mm.down.weight
F16
F16
[4608, 1536]
mm.gate.weight
F16
F16
[1536, 4608]
mm.model.fc.weight
F16
F16
[1536, 1536]
mm.patch_merger.bias
F32
F32
[1536]
mm.patch_merger.weight
F16
F16
[2, 2, 1024, 1536]
mm.post_norm.bias
F32
F32
[1536]
mm.post_norm.weight
F32
F32
[1536]
mm.up.weight
F16
F16
[1536, 4608]
output.weight
BF16
BF16
[1536, 59392]
v.blk.0
v.blk.0.attn_k_norm.weight
F32
F32
[64]
v.blk.0.attn_out.bias
F32
F32
[1024]
v.blk.0.attn_out.weight
F16
F16
[1024, 1024]
v.blk.0.attn_q_norm.weight
F32
F32
[64]
v.blk.0.attn_qkv.bias
F32
F32
[3072]
v.blk.0.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.0.ffn_down.bias
F32
F32
[1024]
v.blk.0.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.0.ffn_gate.bias
F32
F32
[4096]
v.blk.0.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.0.ffn_up.bias
F32
F32
[4096]
v.blk.0.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.0.ln1.weight
F32
F32
[1024]
v.blk.0.ln2.weight
F32
F32
[1024]
v.blk.1
v.blk.1.attn_k_norm.weight
F32
F32
[64]
v.blk.1.attn_out.bias
F32
F32
[1024]
v.blk.1.attn_out.weight
F16
F16
[1024, 1024]
v.blk.1.attn_q_norm.weight
F32
F32
[64]
v.blk.1.attn_qkv.bias
F32
F32
[3072]
v.blk.1.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.1.ffn_down.bias
F32
F32
[1024]
v.blk.1.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.1.ffn_gate.bias
F32
F32
[4096]
v.blk.1.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.1.ffn_up.bias
F32
F32
[4096]
v.blk.1.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.1.ln1.weight
F32
F32
[1024]
v.blk.1.ln2.weight
F32
F32
[1024]
v.blk.2
v.blk.2.attn_k_norm.weight
F32
F32
[64]
v.blk.2.attn_out.bias
F32
F32
[1024]
v.blk.2.attn_out.weight
F16
F16
[1024, 1024]
v.blk.2.attn_q_norm.weight
F32
F32
[64]
v.blk.2.attn_qkv.bias
F32
F32
[3072]
v.blk.2.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.2.ffn_down.bias
F32
F32
[1024]
v.blk.2.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.2.ffn_gate.bias
F32
F32
[4096]
v.blk.2.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.2.ffn_up.bias
F32
F32
[4096]
v.blk.2.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.2.ln1.weight
F32
F32
[1024]
v.blk.2.ln2.weight
F32
F32
[1024]
v.blk.3
v.blk.3.attn_k_norm.weight
F32
F32
[64]
v.blk.3.attn_out.bias
F32
F32
[1024]
v.blk.3.attn_out.weight
F16
F16
[1024, 1024]
v.blk.3.attn_q_norm.weight
F32
F32
[64]
v.blk.3.attn_qkv.bias
F32
F32
[3072]
v.blk.3.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.3.ffn_down.bias
F32
F32
[1024]
v.blk.3.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.3.ffn_gate.bias
F32
F32
[4096]
v.blk.3.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.3.ffn_up.bias
F32
F32
[4096]
v.blk.3.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.3.ln1.weight
F32
F32
[1024]
v.blk.3.ln2.weight
F32
F32
[1024]
v.blk.4
v.blk.4.attn_k_norm.weight
F32
F32
[64]
v.blk.4.attn_out.bias
F32
F32
[1024]
v.blk.4.attn_out.weight
F16
F16
[1024, 1024]
v.blk.4.attn_q_norm.weight
F32
F32
[64]
v.blk.4.attn_qkv.bias
F32
F32
[3072]
v.blk.4.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.4.ffn_down.bias
F32
F32
[1024]
v.blk.4.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.4.ffn_gate.bias
F32
F32
[4096]
v.blk.4.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.4.ffn_up.bias
F32
F32
[4096]
v.blk.4.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.4.ln1.weight
F32
F32
[1024]
v.blk.4.ln2.weight
F32
F32
[1024]
v.blk.5
v.blk.5.attn_k_norm.weight
F32
F32
[64]
v.blk.5.attn_out.bias
F32
F32
[1024]
v.blk.5.attn_out.weight
F16
F16
[1024, 1024]
v.blk.5.attn_q_norm.weight
F32
F32
[64]
v.blk.5.attn_qkv.bias
F32
F32
[3072]
v.blk.5.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.5.ffn_down.bias
F32
F32
[1024]
v.blk.5.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.5.ffn_gate.bias
F32
F32
[4096]
v.blk.5.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.5.ffn_up.bias
F32
F32
[4096]
v.blk.5.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.5.ln1.weight
F32
F32
[1024]
v.blk.5.ln2.weight
F32
F32
[1024]
v.blk.6
v.blk.6.attn_k_norm.weight
F32
F32
[64]
v.blk.6.attn_out.bias
F32
F32
[1024]
v.blk.6.attn_out.weight
F16
F16
[1024, 1024]
v.blk.6.attn_q_norm.weight
F32
F32
[64]
v.blk.6.attn_qkv.bias
F32
F32
[3072]
v.blk.6.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.6.ffn_down.bias
F32
F32
[1024]
v.blk.6.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.6.ffn_gate.bias
F32
F32
[4096]
v.blk.6.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.6.ffn_up.bias
F32
F32
[4096]
v.blk.6.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.6.ln1.weight
F32
F32
[1024]
v.blk.6.ln2.weight
F32
F32
[1024]
v.blk.7
v.blk.7.attn_k_norm.weight
F32
F32
[64]
v.blk.7.attn_out.bias
F32
F32
[1024]
v.blk.7.attn_out.weight
F16
F16
[1024, 1024]
v.blk.7.attn_q_norm.weight
F32
F32
[64]
v.blk.7.attn_qkv.bias
F32
F32
[3072]
v.blk.7.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.7.ffn_down.bias
F32
F32
[1024]
v.blk.7.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.7.ffn_gate.bias
F32
F32
[4096]
v.blk.7.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.7.ffn_up.bias
F32
F32
[4096]
v.blk.7.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.7.ln1.weight
F32
F32
[1024]
v.blk.7.ln2.weight
F32
F32
[1024]
v.blk.8
v.blk.8.attn_k_norm.weight
F32
F32
[64]
v.blk.8.attn_out.bias
F32
F32
[1024]
v.blk.8.attn_out.weight
F16
F16
[1024, 1024]
v.blk.8.attn_q_norm.weight
F32
F32
[64]
v.blk.8.attn_qkv.bias
F32
F32
[3072]
v.blk.8.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.8.ffn_down.bias
F32
F32
[1024]
v.blk.8.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.8.ffn_gate.bias
F32
F32
[4096]
v.blk.8.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.8.ffn_up.bias
F32
F32
[4096]
v.blk.8.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.8.ln1.weight
F32
F32
[1024]
v.blk.8.ln2.weight
F32
F32
[1024]
v.blk.9
v.blk.9.attn_k_norm.weight
F32
F32
[64]
v.blk.9.attn_out.bias
F32
F32
[1024]
v.blk.9.attn_out.weight
F16
F16
[1024, 1024]
v.blk.9.attn_q_norm.weight
F32
F32
[64]
v.blk.9.attn_qkv.bias
F32
F32
[3072]
v.blk.9.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.9.ffn_down.bias
F32
F32
[1024]
v.blk.9.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.9.ffn_gate.bias
F32
F32
[4096]
v.blk.9.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.9.ffn_up.bias
F32
F32
[4096]
v.blk.9.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.9.ln1.weight
F32
F32
[1024]
v.blk.9.ln2.weight
F32
F32
[1024]
v.blk.10
v.blk.10.attn_k_norm.weight
F32
F32
[64]
v.blk.10.attn_out.bias
F32
F32
[1024]
v.blk.10.attn_out.weight
F16
F16
[1024, 1024]
v.blk.10.attn_q_norm.weight
F32
F32
[64]
v.blk.10.attn_qkv.bias
F32
F32
[3072]
v.blk.10.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.10.ffn_down.bias
F32
F32
[1024]
v.blk.10.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.10.ffn_gate.bias
F32
F32
[4096]
v.blk.10.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.10.ffn_up.bias
F32
F32
[4096]
v.blk.10.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.10.ln1.weight
F32
F32
[1024]
v.blk.10.ln2.weight
F32
F32
[1024]
v.blk.11
v.blk.11.attn_k_norm.weight
F32
F32
[64]
v.blk.11.attn_out.bias
F32
F32
[1024]
v.blk.11.attn_out.weight
F16
F16
[1024, 1024]
v.blk.11.attn_q_norm.weight
F32
F32
[64]
v.blk.11.attn_qkv.bias
F32
F32
[3072]
v.blk.11.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.11.ffn_down.bias
F32
F32
[1024]
v.blk.11.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.11.ffn_gate.bias
F32
F32
[4096]
v.blk.11.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.11.ffn_up.bias
F32
F32
[4096]
v.blk.11.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.11.ln1.weight
F32
F32
[1024]
v.blk.11.ln2.weight
F32
F32
[1024]
v.blk.12
v.blk.12.attn_k_norm.weight
F32
F32
[64]
v.blk.12.attn_out.bias
F32
F32
[1024]
v.blk.12.attn_out.weight
F16
F16
[1024, 1024]
v.blk.12.attn_q_norm.weight
F32
F32
[64]
v.blk.12.attn_qkv.bias
F32
F32
[3072]
v.blk.12.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.12.ffn_down.bias
F32
F32
[1024]
v.blk.12.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.12.ffn_gate.bias
F32
F32
[4096]
v.blk.12.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.12.ffn_up.bias
F32
F32
[4096]
v.blk.12.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.12.ln1.weight
F32
F32
[1024]
v.blk.12.ln2.weight
F32
F32
[1024]
v.blk.13
v.blk.13.attn_k_norm.weight
F32
F32
[64]
v.blk.13.attn_out.bias
F32
F32
[1024]
v.blk.13.attn_out.weight
F16
F16
[1024, 1024]
v.blk.13.attn_q_norm.weight
F32
F32
[64]
v.blk.13.attn_qkv.bias
F32
F32
[3072]
v.blk.13.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.13.ffn_down.bias
F32
F32
[1024]
v.blk.13.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.13.ffn_gate.bias
F32
F32
[4096]
v.blk.13.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.13.ffn_up.bias
F32
F32
[4096]
v.blk.13.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.13.ln1.weight
F32
F32
[1024]
v.blk.13.ln2.weight
F32
F32
[1024]
v.blk.14
v.blk.14.attn_k_norm.weight
F32
F32
[64]
v.blk.14.attn_out.bias
F32
F32
[1024]
v.blk.14.attn_out.weight
F16
F16
[1024, 1024]
v.blk.14.attn_q_norm.weight
F32
F32
[64]
v.blk.14.attn_qkv.bias
F32
F32
[3072]
v.blk.14.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.14.ffn_down.bias
F32
F32
[1024]
v.blk.14.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.14.ffn_gate.bias
F32
F32
[4096]
v.blk.14.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.14.ffn_up.bias
F32
F32
[4096]
v.blk.14.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.14.ln1.weight
F32
F32
[1024]
v.blk.14.ln2.weight
F32
F32
[1024]
v.blk.15
v.blk.15.attn_k_norm.weight
F32
F32
[64]
v.blk.15.attn_out.bias
F32
F32
[1024]
v.blk.15.attn_out.weight
F16
F16
[1024, 1024]
v.blk.15.attn_q_norm.weight
F32
F32
[64]
v.blk.15.attn_qkv.bias
F32
F32
[3072]
v.blk.15.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.15.ffn_down.bias
F32
F32
[1024]
v.blk.15.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.15.ffn_gate.bias
F32
F32
[4096]
v.blk.15.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.15.ffn_up.bias
F32
F32
[4096]
v.blk.15.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.15.ln1.weight
F32
F32
[1024]
v.blk.15.ln2.weight
F32
F32
[1024]
v.blk.16
v.blk.16.attn_k_norm.weight
F32
F32
[64]
v.blk.16.attn_out.bias
F32
F32
[1024]
v.blk.16.attn_out.weight
F16
F16
[1024, 1024]
v.blk.16.attn_q_norm.weight
F32
F32
[64]
v.blk.16.attn_qkv.bias
F32
F32
[3072]
v.blk.16.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.16.ffn_down.bias
F32
F32
[1024]
v.blk.16.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.16.ffn_gate.bias
F32
F32
[4096]
v.blk.16.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.16.ffn_up.bias
F32
F32
[4096]
v.blk.16.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.16.ln1.weight
F32
F32
[1024]
v.blk.16.ln2.weight
F32
F32
[1024]
v.blk.17
v.blk.17.attn_k_norm.weight
F32
F32
[64]
v.blk.17.attn_out.bias
F32
F32
[1024]
v.blk.17.attn_out.weight
F16
F16
[1024, 1024]
v.blk.17.attn_q_norm.weight
F32
F32
[64]
v.blk.17.attn_qkv.bias
F32
F32
[3072]
v.blk.17.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.17.ffn_down.bias
F32
F32
[1024]
v.blk.17.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.17.ffn_gate.bias
F32
F32
[4096]
v.blk.17.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.17.ffn_up.bias
F32
F32
[4096]
v.blk.17.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.17.ln1.weight
F32
F32
[1024]
v.blk.17.ln2.weight
F32
F32
[1024]
v.blk.18
v.blk.18.attn_k_norm.weight
F32
F32
[64]
v.blk.18.attn_out.bias
F32
F32
[1024]
v.blk.18.attn_out.weight
F16
F16
[1024, 1024]
v.blk.18.attn_q_norm.weight
F32
F32
[64]
v.blk.18.attn_qkv.bias
F32
F32
[3072]
v.blk.18.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.18.ffn_down.bias
F32
F32
[1024]
v.blk.18.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.18.ffn_gate.bias
F32
F32
[4096]
v.blk.18.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.18.ffn_up.bias
F32
F32
[4096]
v.blk.18.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.18.ln1.weight
F32
F32
[1024]
v.blk.18.ln2.weight
F32
F32
[1024]
v.blk.19
v.blk.19.attn_k_norm.weight
F32
F32
[64]
v.blk.19.attn_out.bias
F32
F32
[1024]
v.blk.19.attn_out.weight
F16
F16
[1024, 1024]
v.blk.19.attn_q_norm.weight
F32
F32
[64]
v.blk.19.attn_qkv.bias
F32
F32
[3072]
v.blk.19.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.19.ffn_down.bias
F32
F32
[1024]
v.blk.19.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.19.ffn_gate.bias
F32
F32
[4096]
v.blk.19.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.19.ffn_up.bias
F32
F32
[4096]
v.blk.19.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.19.ln1.weight
F32
F32
[1024]
v.blk.19.ln2.weight
F32
F32
[1024]
v.blk.20
v.blk.20.attn_k_norm.weight
F32
F32
[64]
v.blk.20.attn_out.bias
F32
F32
[1024]
v.blk.20.attn_out.weight
F16
F16
[1024, 1024]
v.blk.20.attn_q_norm.weight
F32
F32
[64]
v.blk.20.attn_qkv.bias
F32
F32
[3072]
v.blk.20.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.20.ffn_down.bias
F32
F32
[1024]
v.blk.20.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.20.ffn_gate.bias
F32
F32
[4096]
v.blk.20.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.20.ffn_up.bias
F32
F32
[4096]
v.blk.20.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.20.ln1.weight
F32
F32
[1024]
v.blk.20.ln2.weight
F32
F32
[1024]
v.blk.21
v.blk.21.attn_k_norm.weight
F32
F32
[64]
v.blk.21.attn_out.bias
F32
F32
[1024]
v.blk.21.attn_out.weight
F16
F16
[1024, 1024]
v.blk.21.attn_q_norm.weight
F32
F32
[64]
v.blk.21.attn_qkv.bias
F32
F32
[3072]
v.blk.21.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.21.ffn_down.bias
F32
F32
[1024]
v.blk.21.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.21.ffn_gate.bias
F32
F32
[4096]
v.blk.21.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.21.ffn_up.bias
F32
F32
[4096]
v.blk.21.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.21.ln1.weight
F32
F32
[1024]
v.blk.21.ln2.weight
F32
F32
[1024]
v.blk.22
v.blk.22.attn_k_norm.weight
F32
F32
[64]
v.blk.22.attn_out.bias
F32
F32
[1024]
v.blk.22.attn_out.weight
F16
F16
[1024, 1024]
v.blk.22.attn_q_norm.weight
F32
F32
[64]
v.blk.22.attn_qkv.bias
F32
F32
[3072]
v.blk.22.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.22.ffn_down.bias
F32
F32
[1024]
v.blk.22.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.22.ffn_gate.bias
F32
F32
[4096]
v.blk.22.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.22.ffn_up.bias
F32
F32
[4096]
v.blk.22.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.22.ln1.weight
F32
F32
[1024]
v.blk.22.ln2.weight
F32
F32
[1024]
v.blk.23
v.blk.23.attn_k_norm.weight
F32
F32
[64]
v.blk.23.attn_out.bias
F32
F32
[1024]
v.blk.23.attn_out.weight
F16
F16
[1024, 1024]
v.blk.23.attn_q_norm.weight
F32
F32
[64]
v.blk.23.attn_qkv.bias
F32
F32
[3072]
v.blk.23.attn_qkv.weight
F16
F16
[1024, 3072]
v.blk.23.ffn_down.bias
F32
F32
[1024]
v.blk.23.ffn_down.weight
F16
F16
[4096, 1024]
v.blk.23.ffn_gate.bias
F32
F32
[4096]
v.blk.23.ffn_gate.weight
F16
F16
[1024, 4096]
v.blk.23.ffn_up.bias
F32
F32
[4096]
v.blk.23.ffn_up.weight
F16
F16
[1024, 4096]
v.blk.23.ln1.weight
F32
F32
[1024]
v.blk.23.ln2.weight
F32
F32
[1024]
v.patch_embd.bias
F32
F32
[1024]
v.patch_embd_0.weight
F16
F16
[14, 14, 3, 1024]
v.patch_embd_1.weight
F16
F16
[14, 14, 3, 1024]
v.post_ln.weight
F32
F32
[1024]
output_norm.weight
F32
F32
[1536]