Models
Docs
Pricing
Sign in
Download
Models
Download
Docs
Pricing
Sign in
kavai
/
qwen3.5-GPT5
:2b
962
Downloads
Updated
3 days ago
Qwen 3.5 is a family of open-source models that delivers exceptional utility and performance for tool calling and Agentic abilities. Smaller Models may suffer from slower speeds.
Qwen 3.5 is a family of open-source models that delivers exceptional utility and performance for tool calling and Agentic abilities. Smaller Models may suffer from slower speeds.
Cancel
vision
tools
thinking
0.8b
2b
4b
9b
27b
35b
122b
qwen3.5-GPT5:2b
...
/
model
b709d81508a0 · 2.7GB
Metadata
general.architecture
qwen35
qwen35
general.file_type
Q8_0
Q8_0
qwen35.attention.head_count
8
8
qwen35.attention.head_count_kv
[0, 0, 0, 2, 0, ...]
[0, 0, 0, 2, 0, ...]
qwen35.attention.key_length
256
256
qwen35.attention.layer_norm_rms_epsilon
1e-06
1e-06
qwen35.attention.value_length
256
256
qwen35.block_count
24
24
qwen35.context_length
262144
262144
qwen35.embedding_length
2048
2048
qwen35.feed_forward_length
6144
6144
qwen35.full_attention_interval
4
4
qwen35.image_token_id
248056
248056
qwen35.mrope_sections
[11, 11, 10]
[11, 11, 10]
qwen35.rope.dimension_count
64
64
qwen35.rope.dimension_sections
[11, 11, 10]
[11, 11, 10]
qwen35.rope.freq_base
1e+07
1e+07
qwen35.rope.mrope_interleaved
true
true
qwen35.rope.mrope_section
[11, 11, 10]
[11, 11, 10]
qwen35.ssm.conv_kernel
4
4
qwen35.ssm.group_count
16
16
qwen35.ssm.inner_size
2048
2048
qwen35.ssm.state_size
128
128
qwen35.ssm.time_step_rank
16
16
qwen35.ssm.v_head_reordered
true
true
qwen35.vision.attention.head_count
16
16
qwen35.vision.block_count
24
24
qwen35.vision.deepstack_visual_indexes
[]
[]
qwen35.vision.embedding_length
1024
1024
qwen35.vision.image_mean
[0.5, 0.5, 0.5]
[0.5, 0.5, 0.5]
qwen35.vision.image_std
[0.5, 0.5, 0.5]
[0.5, 0.5, 0.5]
qwen35.vision.longest_edge
16777216
16777216
qwen35.vision.num_channels
3
3
qwen35.vision.patch_size
16
16
qwen35.vision.shortest_edge
65536
65536
qwen35.vision.spatial_merge_size
2
2
qwen35.vision.temporal_patch_size
2
2
qwen35.vision_end_token_id
248054
248054
qwen35.vision_start_token_id
248053
248053
tokenizer.ggml.add_eos_token
false
false
tokenizer.ggml.add_padding_token
false
false
tokenizer.ggml.eos_token_id
248046
248046
tokenizer.ggml.merges
[Ġ Ġ, ĠĠ ĠĠ, i n, Ġ t, ĠĠĠĠ ĠĠĠĠ, ...]
[Ġ Ġ, ĠĠ ĠĠ, i n, Ġ t, ĠĠĠĠ ĠĠĠĠ, ...]
tokenizer.ggml.model
gpt2
gpt2
tokenizer.ggml.padding_token_id
248044
248044
tokenizer.ggml.pre
qwen35
qwen35
tokenizer.ggml.scores
[0, 1, 2, 3, 4, ...]
[0, 1, 2, 3, 4, ...]
tokenizer.ggml.token_type
[1, 1, 1, 1, 1, ...]
[1, 1, 1, 1, 1, ...]
tokenizer.ggml.tokens
[!, ", #, $, %, ...]
[!, ", #, $, %, ...]
Tensor
Name
Type
Shape
token_embd.weight
Q8_0
Q8_0
[2048, 248320]
blk.0
blk.0.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.0.attn_norm.weight
F32
F32
[2048]
blk.0.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.0.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.0.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.0.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.0.post_attention_norm.weight
F32
F32
[2048]
blk.0.ssm_a
F32
F32
[16]
blk.0.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.0.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.0.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.0.ssm_dt
F32
F32
[16]
blk.0.ssm_norm.weight
F32
F32
[128]
blk.0.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.1
blk.1.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.1.attn_norm.weight
F32
F32
[2048]
blk.1.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.1.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.1.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.1.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.1.post_attention_norm.weight
F32
F32
[2048]
blk.1.ssm_a
F32
F32
[16]
blk.1.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.1.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.1.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.1.ssm_dt
F32
F32
[16]
blk.1.ssm_norm.weight
F32
F32
[128]
blk.1.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.2
blk.2.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.2.attn_norm.weight
F32
F32
[2048]
blk.2.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.2.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.2.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.2.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.2.post_attention_norm.weight
F32
F32
[2048]
blk.2.ssm_a
F32
F32
[16]
blk.2.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.2.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.2.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.2.ssm_dt
F32
F32
[16]
blk.2.ssm_norm.weight
F32
F32
[128]
blk.2.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.3
blk.3.attn_k.weight
Q8_0
Q8_0
[2048, 512]
blk.3.attn_k_norm.weight
F32
F32
[256]
blk.3.attn_norm.weight
F32
F32
[2048]
blk.3.attn_output.weight
Q8_0
Q8_0
[2048, 2048]
blk.3.attn_q.weight
Q8_0
Q8_0
[2048, 4096]
blk.3.attn_q_norm.weight
F32
F32
[256]
blk.3.attn_v.weight
Q8_0
Q8_0
[2048, 512]
blk.3.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.3.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.3.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.3.post_attention_norm.weight
F32
F32
[2048]
blk.4
blk.4.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.4.attn_norm.weight
F32
F32
[2048]
blk.4.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.4.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.4.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.4.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.4.post_attention_norm.weight
F32
F32
[2048]
blk.4.ssm_a
F32
F32
[16]
blk.4.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.4.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.4.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.4.ssm_dt
F32
F32
[16]
blk.4.ssm_norm.weight
F32
F32
[128]
blk.4.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.5
blk.5.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.5.attn_norm.weight
F32
F32
[2048]
blk.5.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.5.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.5.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.5.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.5.post_attention_norm.weight
F32
F32
[2048]
blk.5.ssm_a
F32
F32
[16]
blk.5.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.5.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.5.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.5.ssm_dt
F32
F32
[16]
blk.5.ssm_norm.weight
F32
F32
[128]
blk.5.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.6
blk.6.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.6.attn_norm.weight
F32
F32
[2048]
blk.6.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.6.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.6.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.6.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.6.post_attention_norm.weight
F32
F32
[2048]
blk.6.ssm_a
F32
F32
[16]
blk.6.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.6.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.6.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.6.ssm_dt
F32
F32
[16]
blk.6.ssm_norm.weight
F32
F32
[128]
blk.6.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.7
blk.7.attn_k.weight
Q8_0
Q8_0
[2048, 512]
blk.7.attn_k_norm.weight
F32
F32
[256]
blk.7.attn_norm.weight
F32
F32
[2048]
blk.7.attn_output.weight
Q8_0
Q8_0
[2048, 2048]
blk.7.attn_q.weight
Q8_0
Q8_0
[2048, 4096]
blk.7.attn_q_norm.weight
F32
F32
[256]
blk.7.attn_v.weight
Q8_0
Q8_0
[2048, 512]
blk.7.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.7.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.7.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.7.post_attention_norm.weight
F32
F32
[2048]
blk.8
blk.8.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.8.attn_norm.weight
F32
F32
[2048]
blk.8.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.8.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.8.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.8.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.8.post_attention_norm.weight
F32
F32
[2048]
blk.8.ssm_a
F32
F32
[16]
blk.8.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.8.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.8.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.8.ssm_dt
F32
F32
[16]
blk.8.ssm_norm.weight
F32
F32
[128]
blk.8.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.9
blk.9.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.9.attn_norm.weight
F32
F32
[2048]
blk.9.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.9.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.9.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.9.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.9.post_attention_norm.weight
F32
F32
[2048]
blk.9.ssm_a
F32
F32
[16]
blk.9.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.9.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.9.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.9.ssm_dt
F32
F32
[16]
blk.9.ssm_norm.weight
F32
F32
[128]
blk.9.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.10
blk.10.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.10.attn_norm.weight
F32
F32
[2048]
blk.10.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.10.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.10.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.10.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.10.post_attention_norm.weight
F32
F32
[2048]
blk.10.ssm_a
F32
F32
[16]
blk.10.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.10.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.10.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.10.ssm_dt
F32
F32
[16]
blk.10.ssm_norm.weight
F32
F32
[128]
blk.10.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.11
blk.11.attn_k.weight
Q8_0
Q8_0
[2048, 512]
blk.11.attn_k_norm.weight
F32
F32
[256]
blk.11.attn_norm.weight
F32
F32
[2048]
blk.11.attn_output.weight
Q8_0
Q8_0
[2048, 2048]
blk.11.attn_q.weight
Q8_0
Q8_0
[2048, 4096]
blk.11.attn_q_norm.weight
F32
F32
[256]
blk.11.attn_v.weight
Q8_0
Q8_0
[2048, 512]
blk.11.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.11.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.11.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.11.post_attention_norm.weight
F32
F32
[2048]
blk.12
blk.12.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.12.attn_norm.weight
F32
F32
[2048]
blk.12.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.12.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.12.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.12.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.12.post_attention_norm.weight
F32
F32
[2048]
blk.12.ssm_a
F32
F32
[16]
blk.12.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.12.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.12.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.12.ssm_dt
F32
F32
[16]
blk.12.ssm_norm.weight
F32
F32
[128]
blk.12.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.13
blk.13.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.13.attn_norm.weight
F32
F32
[2048]
blk.13.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.13.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.13.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.13.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.13.post_attention_norm.weight
F32
F32
[2048]
blk.13.ssm_a
F32
F32
[16]
blk.13.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.13.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.13.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.13.ssm_dt
F32
F32
[16]
blk.13.ssm_norm.weight
F32
F32
[128]
blk.13.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.14
blk.14.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.14.attn_norm.weight
F32
F32
[2048]
blk.14.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.14.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.14.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.14.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.14.post_attention_norm.weight
F32
F32
[2048]
blk.14.ssm_a
F32
F32
[16]
blk.14.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.14.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.14.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.14.ssm_dt
F32
F32
[16]
blk.14.ssm_norm.weight
F32
F32
[128]
blk.14.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.15
blk.15.attn_k.weight
Q8_0
Q8_0
[2048, 512]
blk.15.attn_k_norm.weight
F32
F32
[256]
blk.15.attn_norm.weight
F32
F32
[2048]
blk.15.attn_output.weight
Q8_0
Q8_0
[2048, 2048]
blk.15.attn_q.weight
Q8_0
Q8_0
[2048, 4096]
blk.15.attn_q_norm.weight
F32
F32
[256]
blk.15.attn_v.weight
Q8_0
Q8_0
[2048, 512]
blk.15.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.15.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.15.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.15.post_attention_norm.weight
F32
F32
[2048]
blk.16
blk.16.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.16.attn_norm.weight
F32
F32
[2048]
blk.16.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.16.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.16.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.16.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.16.post_attention_norm.weight
F32
F32
[2048]
blk.16.ssm_a
F32
F32
[16]
blk.16.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.16.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.16.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.16.ssm_dt
F32
F32
[16]
blk.16.ssm_norm.weight
F32
F32
[128]
blk.16.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.17
blk.17.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.17.attn_norm.weight
F32
F32
[2048]
blk.17.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.17.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.17.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.17.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.17.post_attention_norm.weight
F32
F32
[2048]
blk.17.ssm_a
F32
F32
[16]
blk.17.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.17.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.17.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.17.ssm_dt
F32
F32
[16]
blk.17.ssm_norm.weight
F32
F32
[128]
blk.17.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.18
blk.18.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.18.attn_norm.weight
F32
F32
[2048]
blk.18.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.18.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.18.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.18.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.18.post_attention_norm.weight
F32
F32
[2048]
blk.18.ssm_a
F32
F32
[16]
blk.18.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.18.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.18.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.18.ssm_dt
F32
F32
[16]
blk.18.ssm_norm.weight
F32
F32
[128]
blk.18.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.19
blk.19.attn_k.weight
Q8_0
Q8_0
[2048, 512]
blk.19.attn_k_norm.weight
F32
F32
[256]
blk.19.attn_norm.weight
F32
F32
[2048]
blk.19.attn_output.weight
Q8_0
Q8_0
[2048, 2048]
blk.19.attn_q.weight
Q8_0
Q8_0
[2048, 4096]
blk.19.attn_q_norm.weight
F32
F32
[256]
blk.19.attn_v.weight
Q8_0
Q8_0
[2048, 512]
blk.19.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.19.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.19.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.19.post_attention_norm.weight
F32
F32
[2048]
blk.20
blk.20.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.20.attn_norm.weight
F32
F32
[2048]
blk.20.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.20.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.20.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.20.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.20.post_attention_norm.weight
F32
F32
[2048]
blk.20.ssm_a
F32
F32
[16]
blk.20.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.20.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.20.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.20.ssm_dt
F32
F32
[16]
blk.20.ssm_norm.weight
F32
F32
[128]
blk.20.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.21
blk.21.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.21.attn_norm.weight
F32
F32
[2048]
blk.21.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.21.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.21.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.21.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.21.post_attention_norm.weight
F32
F32
[2048]
blk.21.ssm_a
F32
F32
[16]
blk.21.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.21.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.21.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.21.ssm_dt
F32
F32
[16]
blk.21.ssm_norm.weight
F32
F32
[128]
blk.21.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.22
blk.22.attn_gate.weight
Q8_0
Q8_0
[2048, 2048]
blk.22.attn_norm.weight
F32
F32
[2048]
blk.22.attn_qkv.weight
Q8_0
Q8_0
[2048, 6144]
blk.22.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.22.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.22.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.22.post_attention_norm.weight
F32
F32
[2048]
blk.22.ssm_a
F32
F32
[16]
blk.22.ssm_alpha.weight
Q8_0
Q8_0
[2048, 16]
blk.22.ssm_beta.weight
Q8_0
Q8_0
[2048, 16]
blk.22.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.22.ssm_dt
F32
F32
[16]
blk.22.ssm_norm.weight
F32
F32
[128]
blk.22.ssm_out.weight
Q8_0
Q8_0
[2048, 2048]
blk.23
blk.23.attn_k.weight
Q8_0
Q8_0
[2048, 512]
blk.23.attn_k_norm.weight
F32
F32
[256]
blk.23.attn_norm.weight
F32
F32
[2048]
blk.23.attn_output.weight
Q8_0
Q8_0
[2048, 2048]
blk.23.attn_q.weight
Q8_0
Q8_0
[2048, 4096]
blk.23.attn_q_norm.weight
F32
F32
[256]
blk.23.attn_v.weight
Q8_0
Q8_0
[2048, 512]
blk.23.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
blk.23.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
blk.23.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
blk.23.post_attention_norm.weight
F32
F32
[2048]
mtp.fc.weight
Q8_0
Q8_0
[4096, 2048]
mtp.layers.0.attn_k.weight
Q8_0
Q8_0
[2048, 512]
mtp.layers.0.attn_k_norm.weight
F32
F32
[256]
mtp.layers.0.attn_norm.weight
F32
F32
[2048]
mtp.layers.0.attn_output.weight
Q8_0
Q8_0
[2048, 2048]
mtp.layers.0.attn_q.weight
Q8_0
Q8_0
[2048, 4096]
mtp.layers.0.attn_q_norm.weight
F32
F32
[256]
mtp.layers.0.attn_v.weight
Q8_0
Q8_0
[2048, 512]
mtp.layers.0.ffn_down.weight
Q8_0
Q8_0
[6144, 2048]
mtp.layers.0.ffn_gate.weight
Q8_0
Q8_0
[2048, 6144]
mtp.layers.0.ffn_up.weight
Q8_0
Q8_0
[2048, 6144]
mtp.layers.0.post_attention_norm.weight
F32
F32
[2048]
mtp.norm.weight
F32
F32
[2048]
mtp.pre_fc_norm_embedding.weight
F32
F32
[2048]
mtp.pre_fc_norm_hidden.weight
F32
F32
[2048]
v.blk.0
v.blk.0.attn_k.bias
F32
F32
[1024]
v.blk.0.attn_k.weight
F16
F16
[1024, 1024]
v.blk.0.attn_out.bias
F32
F32
[1024]
v.blk.0.attn_out.weight
F16
F16
[1024, 1024]
v.blk.0.attn_q.bias
F32
F32
[1024]
v.blk.0.attn_q.weight
F16
F16
[1024, 1024]
v.blk.0.attn_v.bias
F32
F32
[1024]
v.blk.0.attn_v.weight
F16
F16
[1024, 1024]
v.blk.0.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.0.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.0.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.0.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.0.norm1.bias
F32
F32
[1024]
v.blk.0.norm1.weight
F32
F32
[1024]
v.blk.0.norm2.bias
F32
F32
[1024]
v.blk.0.norm2.weight
F32
F32
[1024]
v.blk.1
v.blk.1.attn_k.bias
F32
F32
[1024]
v.blk.1.attn_k.weight
F16
F16
[1024, 1024]
v.blk.1.attn_out.bias
F32
F32
[1024]
v.blk.1.attn_out.weight
F16
F16
[1024, 1024]
v.blk.1.attn_q.bias
F32
F32
[1024]
v.blk.1.attn_q.weight
F16
F16
[1024, 1024]
v.blk.1.attn_v.bias
F32
F32
[1024]
v.blk.1.attn_v.weight
F16
F16
[1024, 1024]
v.blk.1.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.1.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.1.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.1.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.1.norm1.bias
F32
F32
[1024]
v.blk.1.norm1.weight
F32
F32
[1024]
v.blk.1.norm2.bias
F32
F32
[1024]
v.blk.1.norm2.weight
F32
F32
[1024]
v.blk.2
v.blk.2.attn_k.bias
F32
F32
[1024]
v.blk.2.attn_k.weight
F16
F16
[1024, 1024]
v.blk.2.attn_out.bias
F32
F32
[1024]
v.blk.2.attn_out.weight
F16
F16
[1024, 1024]
v.blk.2.attn_q.bias
F32
F32
[1024]
v.blk.2.attn_q.weight
F16
F16
[1024, 1024]
v.blk.2.attn_v.bias
F32
F32
[1024]
v.blk.2.attn_v.weight
F16
F16
[1024, 1024]
v.blk.2.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.2.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.2.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.2.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.2.norm1.bias
F32
F32
[1024]
v.blk.2.norm1.weight
F32
F32
[1024]
v.blk.2.norm2.bias
F32
F32
[1024]
v.blk.2.norm2.weight
F32
F32
[1024]
v.blk.3
v.blk.3.attn_k.bias
F32
F32
[1024]
v.blk.3.attn_k.weight
F16
F16
[1024, 1024]
v.blk.3.attn_out.bias
F32
F32
[1024]
v.blk.3.attn_out.weight
F16
F16
[1024, 1024]
v.blk.3.attn_q.bias
F32
F32
[1024]
v.blk.3.attn_q.weight
F16
F16
[1024, 1024]
v.blk.3.attn_v.bias
F32
F32
[1024]
v.blk.3.attn_v.weight
F16
F16
[1024, 1024]
v.blk.3.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.3.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.3.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.3.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.3.norm1.bias
F32
F32
[1024]
v.blk.3.norm1.weight
F32
F32
[1024]
v.blk.3.norm2.bias
F32
F32
[1024]
v.blk.3.norm2.weight
F32
F32
[1024]
v.blk.4
v.blk.4.attn_k.bias
F32
F32
[1024]
v.blk.4.attn_k.weight
F16
F16
[1024, 1024]
v.blk.4.attn_out.bias
F32
F32
[1024]
v.blk.4.attn_out.weight
F16
F16
[1024, 1024]
v.blk.4.attn_q.bias
F32
F32
[1024]
v.blk.4.attn_q.weight
F16
F16
[1024, 1024]
v.blk.4.attn_v.bias
F32
F32
[1024]
v.blk.4.attn_v.weight
F16
F16
[1024, 1024]
v.blk.4.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.4.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.4.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.4.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.4.norm1.bias
F32
F32
[1024]
v.blk.4.norm1.weight
F32
F32
[1024]
v.blk.4.norm2.bias
F32
F32
[1024]
v.blk.4.norm2.weight
F32
F32
[1024]
v.blk.5
v.blk.5.attn_k.bias
F32
F32
[1024]
v.blk.5.attn_k.weight
F16
F16
[1024, 1024]
v.blk.5.attn_out.bias
F32
F32
[1024]
v.blk.5.attn_out.weight
F16
F16
[1024, 1024]
v.blk.5.attn_q.bias
F32
F32
[1024]
v.blk.5.attn_q.weight
F16
F16
[1024, 1024]
v.blk.5.attn_v.bias
F32
F32
[1024]
v.blk.5.attn_v.weight
F16
F16
[1024, 1024]
v.blk.5.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.5.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.5.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.5.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.5.norm1.bias
F32
F32
[1024]
v.blk.5.norm1.weight
F32
F32
[1024]
v.blk.5.norm2.bias
F32
F32
[1024]
v.blk.5.norm2.weight
F32
F32
[1024]
v.blk.6
v.blk.6.attn_k.bias
F32
F32
[1024]
v.blk.6.attn_k.weight
F16
F16
[1024, 1024]
v.blk.6.attn_out.bias
F32
F32
[1024]
v.blk.6.attn_out.weight
F16
F16
[1024, 1024]
v.blk.6.attn_q.bias
F32
F32
[1024]
v.blk.6.attn_q.weight
F16
F16
[1024, 1024]
v.blk.6.attn_v.bias
F32
F32
[1024]
v.blk.6.attn_v.weight
F16
F16
[1024, 1024]
v.blk.6.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.6.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.6.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.6.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.6.norm1.bias
F32
F32
[1024]
v.blk.6.norm1.weight
F32
F32
[1024]
v.blk.6.norm2.bias
F32
F32
[1024]
v.blk.6.norm2.weight
F32
F32
[1024]
v.blk.7
v.blk.7.attn_k.bias
F32
F32
[1024]
v.blk.7.attn_k.weight
F16
F16
[1024, 1024]
v.blk.7.attn_out.bias
F32
F32
[1024]
v.blk.7.attn_out.weight
F16
F16
[1024, 1024]
v.blk.7.attn_q.bias
F32
F32
[1024]
v.blk.7.attn_q.weight
F16
F16
[1024, 1024]
v.blk.7.attn_v.bias
F32
F32
[1024]
v.blk.7.attn_v.weight
F16
F16
[1024, 1024]
v.blk.7.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.7.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.7.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.7.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.7.norm1.bias
F32
F32
[1024]
v.blk.7.norm1.weight
F32
F32
[1024]
v.blk.7.norm2.bias
F32
F32
[1024]
v.blk.7.norm2.weight
F32
F32
[1024]
v.blk.8
v.blk.8.attn_k.bias
F32
F32
[1024]
v.blk.8.attn_k.weight
F16
F16
[1024, 1024]
v.blk.8.attn_out.bias
F32
F32
[1024]
v.blk.8.attn_out.weight
F16
F16
[1024, 1024]
v.blk.8.attn_q.bias
F32
F32
[1024]
v.blk.8.attn_q.weight
F16
F16
[1024, 1024]
v.blk.8.attn_v.bias
F32
F32
[1024]
v.blk.8.attn_v.weight
F16
F16
[1024, 1024]
v.blk.8.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.8.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.8.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.8.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.8.norm1.bias
F32
F32
[1024]
v.blk.8.norm1.weight
F32
F32
[1024]
v.blk.8.norm2.bias
F32
F32
[1024]
v.blk.8.norm2.weight
F32
F32
[1024]
v.blk.9
v.blk.9.attn_k.bias
F32
F32
[1024]
v.blk.9.attn_k.weight
F16
F16
[1024, 1024]
v.blk.9.attn_out.bias
F32
F32
[1024]
v.blk.9.attn_out.weight
F16
F16
[1024, 1024]
v.blk.9.attn_q.bias
F32
F32
[1024]
v.blk.9.attn_q.weight
F16
F16
[1024, 1024]
v.blk.9.attn_v.bias
F32
F32
[1024]
v.blk.9.attn_v.weight
F16
F16
[1024, 1024]
v.blk.9.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.9.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.9.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.9.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.9.norm1.bias
F32
F32
[1024]
v.blk.9.norm1.weight
F32
F32
[1024]
v.blk.9.norm2.bias
F32
F32
[1024]
v.blk.9.norm2.weight
F32
F32
[1024]
v.blk.10
v.blk.10.attn_k.bias
F32
F32
[1024]
v.blk.10.attn_k.weight
F16
F16
[1024, 1024]
v.blk.10.attn_out.bias
F32
F32
[1024]
v.blk.10.attn_out.weight
F16
F16
[1024, 1024]
v.blk.10.attn_q.bias
F32
F32
[1024]
v.blk.10.attn_q.weight
F16
F16
[1024, 1024]
v.blk.10.attn_v.bias
F32
F32
[1024]
v.blk.10.attn_v.weight
F16
F16
[1024, 1024]
v.blk.10.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.10.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.10.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.10.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.10.norm1.bias
F32
F32
[1024]
v.blk.10.norm1.weight
F32
F32
[1024]
v.blk.10.norm2.bias
F32
F32
[1024]
v.blk.10.norm2.weight
F32
F32
[1024]
v.blk.11
v.blk.11.attn_k.bias
F32
F32
[1024]
v.blk.11.attn_k.weight
F16
F16
[1024, 1024]
v.blk.11.attn_out.bias
F32
F32
[1024]
v.blk.11.attn_out.weight
F16
F16
[1024, 1024]
v.blk.11.attn_q.bias
F32
F32
[1024]
v.blk.11.attn_q.weight
F16
F16
[1024, 1024]
v.blk.11.attn_v.bias
F32
F32
[1024]
v.blk.11.attn_v.weight
F16
F16
[1024, 1024]
v.blk.11.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.11.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.11.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.11.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.11.norm1.bias
F32
F32
[1024]
v.blk.11.norm1.weight
F32
F32
[1024]
v.blk.11.norm2.bias
F32
F32
[1024]
v.blk.11.norm2.weight
F32
F32
[1024]
v.blk.12
v.blk.12.attn_k.bias
F32
F32
[1024]
v.blk.12.attn_k.weight
F16
F16
[1024, 1024]
v.blk.12.attn_out.bias
F32
F32
[1024]
v.blk.12.attn_out.weight
F16
F16
[1024, 1024]
v.blk.12.attn_q.bias
F32
F32
[1024]
v.blk.12.attn_q.weight
F16
F16
[1024, 1024]
v.blk.12.attn_v.bias
F32
F32
[1024]
v.blk.12.attn_v.weight
F16
F16
[1024, 1024]
v.blk.12.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.12.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.12.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.12.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.12.norm1.bias
F32
F32
[1024]
v.blk.12.norm1.weight
F32
F32
[1024]
v.blk.12.norm2.bias
F32
F32
[1024]
v.blk.12.norm2.weight
F32
F32
[1024]
v.blk.13
v.blk.13.attn_k.bias
F32
F32
[1024]
v.blk.13.attn_k.weight
F16
F16
[1024, 1024]
v.blk.13.attn_out.bias
F32
F32
[1024]
v.blk.13.attn_out.weight
F16
F16
[1024, 1024]
v.blk.13.attn_q.bias
F32
F32
[1024]
v.blk.13.attn_q.weight
F16
F16
[1024, 1024]
v.blk.13.attn_v.bias
F32
F32
[1024]
v.blk.13.attn_v.weight
F16
F16
[1024, 1024]
v.blk.13.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.13.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.13.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.13.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.13.norm1.bias
F32
F32
[1024]
v.blk.13.norm1.weight
F32
F32
[1024]
v.blk.13.norm2.bias
F32
F32
[1024]
v.blk.13.norm2.weight
F32
F32
[1024]
v.blk.14
v.blk.14.attn_k.bias
F32
F32
[1024]
v.blk.14.attn_k.weight
F16
F16
[1024, 1024]
v.blk.14.attn_out.bias
F32
F32
[1024]
v.blk.14.attn_out.weight
F16
F16
[1024, 1024]
v.blk.14.attn_q.bias
F32
F32
[1024]
v.blk.14.attn_q.weight
F16
F16
[1024, 1024]
v.blk.14.attn_v.bias
F32
F32
[1024]
v.blk.14.attn_v.weight
F16
F16
[1024, 1024]
v.blk.14.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.14.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.14.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.14.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.14.norm1.bias
F32
F32
[1024]
v.blk.14.norm1.weight
F32
F32
[1024]
v.blk.14.norm2.bias
F32
F32
[1024]
v.blk.14.norm2.weight
F32
F32
[1024]
v.blk.15
v.blk.15.attn_k.bias
F32
F32
[1024]
v.blk.15.attn_k.weight
F16
F16
[1024, 1024]
v.blk.15.attn_out.bias
F32
F32
[1024]
v.blk.15.attn_out.weight
F16
F16
[1024, 1024]
v.blk.15.attn_q.bias
F32
F32
[1024]
v.blk.15.attn_q.weight
F16
F16
[1024, 1024]
v.blk.15.attn_v.bias
F32
F32
[1024]
v.blk.15.attn_v.weight
F16
F16
[1024, 1024]
v.blk.15.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.15.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.15.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.15.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.15.norm1.bias
F32
F32
[1024]
v.blk.15.norm1.weight
F32
F32
[1024]
v.blk.15.norm2.bias
F32
F32
[1024]
v.blk.15.norm2.weight
F32
F32
[1024]
v.blk.16
v.blk.16.attn_k.bias
F32
F32
[1024]
v.blk.16.attn_k.weight
F16
F16
[1024, 1024]
v.blk.16.attn_out.bias
F32
F32
[1024]
v.blk.16.attn_out.weight
F16
F16
[1024, 1024]
v.blk.16.attn_q.bias
F32
F32
[1024]
v.blk.16.attn_q.weight
F16
F16
[1024, 1024]
v.blk.16.attn_v.bias
F32
F32
[1024]
v.blk.16.attn_v.weight
F16
F16
[1024, 1024]
v.blk.16.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.16.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.16.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.16.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.16.norm1.bias
F32
F32
[1024]
v.blk.16.norm1.weight
F32
F32
[1024]
v.blk.16.norm2.bias
F32
F32
[1024]
v.blk.16.norm2.weight
F32
F32
[1024]
v.blk.17
v.blk.17.attn_k.bias
F32
F32
[1024]
v.blk.17.attn_k.weight
F16
F16
[1024, 1024]
v.blk.17.attn_out.bias
F32
F32
[1024]
v.blk.17.attn_out.weight
F16
F16
[1024, 1024]
v.blk.17.attn_q.bias
F32
F32
[1024]
v.blk.17.attn_q.weight
F16
F16
[1024, 1024]
v.blk.17.attn_v.bias
F32
F32
[1024]
v.blk.17.attn_v.weight
F16
F16
[1024, 1024]
v.blk.17.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.17.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.17.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.17.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.17.norm1.bias
F32
F32
[1024]
v.blk.17.norm1.weight
F32
F32
[1024]
v.blk.17.norm2.bias
F32
F32
[1024]
v.blk.17.norm2.weight
F32
F32
[1024]
v.blk.18
v.blk.18.attn_k.bias
F32
F32
[1024]
v.blk.18.attn_k.weight
F16
F16
[1024, 1024]
v.blk.18.attn_out.bias
F32
F32
[1024]
v.blk.18.attn_out.weight
F16
F16
[1024, 1024]
v.blk.18.attn_q.bias
F32
F32
[1024]
v.blk.18.attn_q.weight
F16
F16
[1024, 1024]
v.blk.18.attn_v.bias
F32
F32
[1024]
v.blk.18.attn_v.weight
F16
F16
[1024, 1024]
v.blk.18.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.18.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.18.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.18.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.18.norm1.bias
F32
F32
[1024]
v.blk.18.norm1.weight
F32
F32
[1024]
v.blk.18.norm2.bias
F32
F32
[1024]
v.blk.18.norm2.weight
F32
F32
[1024]
v.blk.19
v.blk.19.attn_k.bias
F32
F32
[1024]
v.blk.19.attn_k.weight
F16
F16
[1024, 1024]
v.blk.19.attn_out.bias
F32
F32
[1024]
v.blk.19.attn_out.weight
F16
F16
[1024, 1024]
v.blk.19.attn_q.bias
F32
F32
[1024]
v.blk.19.attn_q.weight
F16
F16
[1024, 1024]
v.blk.19.attn_v.bias
F32
F32
[1024]
v.blk.19.attn_v.weight
F16
F16
[1024, 1024]
v.blk.19.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.19.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.19.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.19.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.19.norm1.bias
F32
F32
[1024]
v.blk.19.norm1.weight
F32
F32
[1024]
v.blk.19.norm2.bias
F32
F32
[1024]
v.blk.19.norm2.weight
F32
F32
[1024]
v.blk.20
v.blk.20.attn_k.bias
F32
F32
[1024]
v.blk.20.attn_k.weight
F16
F16
[1024, 1024]
v.blk.20.attn_out.bias
F32
F32
[1024]
v.blk.20.attn_out.weight
F16
F16
[1024, 1024]
v.blk.20.attn_q.bias
F32
F32
[1024]
v.blk.20.attn_q.weight
F16
F16
[1024, 1024]
v.blk.20.attn_v.bias
F32
F32
[1024]
v.blk.20.attn_v.weight
F16
F16
[1024, 1024]
v.blk.20.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.20.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.20.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.20.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.20.norm1.bias
F32
F32
[1024]
v.blk.20.norm1.weight
F32
F32
[1024]
v.blk.20.norm2.bias
F32
F32
[1024]
v.blk.20.norm2.weight
F32
F32
[1024]
v.blk.21
v.blk.21.attn_k.bias
F32
F32
[1024]
v.blk.21.attn_k.weight
F16
F16
[1024, 1024]
v.blk.21.attn_out.bias
F32
F32
[1024]
v.blk.21.attn_out.weight
F16
F16
[1024, 1024]
v.blk.21.attn_q.bias
F32
F32
[1024]
v.blk.21.attn_q.weight
F16
F16
[1024, 1024]
v.blk.21.attn_v.bias
F32
F32
[1024]
v.blk.21.attn_v.weight
F16
F16
[1024, 1024]
v.blk.21.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.21.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.21.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.21.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.21.norm1.bias
F32
F32
[1024]
v.blk.21.norm1.weight
F32
F32
[1024]
v.blk.21.norm2.bias
F32
F32
[1024]
v.blk.21.norm2.weight
F32
F32
[1024]
v.blk.22
v.blk.22.attn_k.bias
F32
F32
[1024]
v.blk.22.attn_k.weight
F16
F16
[1024, 1024]
v.blk.22.attn_out.bias
F32
F32
[1024]
v.blk.22.attn_out.weight
F16
F16
[1024, 1024]
v.blk.22.attn_q.bias
F32
F32
[1024]
v.blk.22.attn_q.weight
F16
F16
[1024, 1024]
v.blk.22.attn_v.bias
F32
F32
[1024]
v.blk.22.attn_v.weight
F16
F16
[1024, 1024]
v.blk.22.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.22.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.22.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.22.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.22.norm1.bias
F32
F32
[1024]
v.blk.22.norm1.weight
F32
F32
[1024]
v.blk.22.norm2.bias
F32
F32
[1024]
v.blk.22.norm2.weight
F32
F32
[1024]
v.blk.23
v.blk.23.attn_k.bias
F32
F32
[1024]
v.blk.23.attn_k.weight
F16
F16
[1024, 1024]
v.blk.23.attn_out.bias
F32
F32
[1024]
v.blk.23.attn_out.weight
F16
F16
[1024, 1024]
v.blk.23.attn_q.bias
F32
F32
[1024]
v.blk.23.attn_q.weight
F16
F16
[1024, 1024]
v.blk.23.attn_v.bias
F32
F32
[1024]
v.blk.23.attn_v.weight
F16
F16
[1024, 1024]
v.blk.23.mlp.linear_fc1.bias
F32
F32
[4096]
v.blk.23.mlp.linear_fc1.weight
F16
F16
[1024, 4096]
v.blk.23.mlp.linear_fc2.bias
F32
F32
[1024]
v.blk.23.mlp.linear_fc2.weight
F16
F16
[4096, 1024]
v.blk.23.norm1.bias
F32
F32
[1024]
v.blk.23.norm1.weight
F32
F32
[1024]
v.blk.23.norm2.bias
F32
F32
[1024]
v.blk.23.norm2.weight
F32
F32
[1024]
v.merger.linear_fc1.bias
F32
F32
[4096]
v.merger.linear_fc1.weight
F16
F16
[4096, 4096]
v.merger.linear_fc2.bias
F32
F32
[2048]
v.merger.linear_fc2.weight
F16
F16
[4096, 2048]
v.merger.norm.bias
F32
F32
[1024]
v.merger.norm.weight
F32
F32
[1024]
v.patch_embed.bias
F32
F32
[1024]
v.patch_embed.weight
F16
F16
[16, 16, 2, 3072]
v.pos_embed.weight
F16
F16
[1024, 2304]
output_norm.weight
F32
F32
[2048]