Models
Docs
Pricing
Sign in
Download
Models
Download
Docs
Pricing
Sign in
nemotron3
:33b-q8
111
Downloads
Updated
20 hours ago
Cancel
vision
tools
thinking
audio
33b
nemotron3:33b-q8
...
/
model
c34ddc773b02 · 36GB
Metadata
general.architecture
nemotron_h_omni
nemotron_h_omni
general.file_type
Q8_0
Q8_0
nemotron_h_omni.attention.head_count
32
32
nemotron_h_omni.attention.head_count_kv
[0, 0, 0, 0, 0, ...]
[0, 0, 0, 0, 0, ...]
nemotron_h_omni.attention.key_length
128
128
nemotron_h_omni.attention.layer_norm_epsilon
1e-05
1e-05
nemotron_h_omni.attention.layer_norm_rms_epsilon
1e-05
1e-05
nemotron_h_omni.attention.value_length
128
128
nemotron_h_omni.audio.attention.head_count
8
8
nemotron_h_omni.audio.attention.layer_norm_epsilon
1e-05
1e-05
nemotron_h_omni.audio.block_count
24
24
nemotron_h_omni.audio.conv_kernel_size
9
9
nemotron_h_omni.audio.embedding_length
1024
1024
nemotron_h_omni.audio.feed_forward_length
4096
4096
nemotron_h_omni.audio.num_mel_bins
128
128
nemotron_h_omni.audio.projection_hidden_size
4096
4096
nemotron_h_omni.audio.sample_rate
16000
16000
nemotron_h_omni.audio.scale_input
false
false
nemotron_h_omni.audio.sound_token_id
27
27
nemotron_h_omni.audio.subsampling_conv_channels
256
256
nemotron_h_omni.audio.subsampling_conv_kernel_size
3
3
nemotron_h_omni.audio.subsampling_conv_stride
2
2
nemotron_h_omni.audio.subsampling_factor
8
8
nemotron_h_omni.block_count
52
52
nemotron_h_omni.context_length
131072
131072
nemotron_h_omni.embedding_length
2688
2688
nemotron_h_omni.expert_count
128
128
nemotron_h_omni.expert_feed_forward_length
1856
1856
nemotron_h_omni.expert_group_count
1
1
nemotron_h_omni.expert_group_used_count
1
1
nemotron_h_omni.expert_shared_count
1
1
nemotron_h_omni.expert_shared_feed_forward_length
3712
3712
nemotron_h_omni.expert_used_count
6
6
nemotron_h_omni.expert_weights_norm
true
true
nemotron_h_omni.expert_weights_scale
2.5
2.5
nemotron_h_omni.feed_forward_length
[0, 1856, 0, 1856, 0, ...]
[0, 1856, 0, 1856, 0, ...]
nemotron_h_omni.rope.dimension_count
128
128
nemotron_h_omni.rope.freq_base
10000
10000
nemotron_h_omni.ssm.conv_kernel
4
4
nemotron_h_omni.ssm.group_count
8
8
nemotron_h_omni.ssm.inner_size
4096
4096
nemotron_h_omni.ssm.state_size
128
128
nemotron_h_omni.ssm.time_step_rank
64
64
nemotron_h_omni.vision.attention.head_count
16
16
nemotron_h_omni.vision.attention.layer_norm_epsilon
1e-06
1e-06
nemotron_h_omni.vision.block_count
32
32
nemotron_h_omni.vision.embedding_length
1280
1280
nemotron_h_omni.vision.feed_forward_length
5120
5120
nemotron_h_omni.vision.image_end_token_id
20
20
nemotron_h_omni.vision.image_mean
[0.48145467, 0.4578275, 0.40821072]
[0.48145467, 0.4578275, 0.40821072]
nemotron_h_omni.vision.image_size
512
512
nemotron_h_omni.vision.image_start_token_id
19
19
nemotron_h_omni.vision.image_std
[0.26862955, 0.2613026, 0.2757771]
[0.26862955, 0.2613026, 0.2757771]
nemotron_h_omni.vision.image_token_id
18
18
nemotron_h_omni.vision.max_num_patches
13312
13312
nemotron_h_omni.vision.max_tiles
12
12
nemotron_h_omni.vision.min_num_patches
1024
1024
nemotron_h_omni.vision.num_channels
3
3
nemotron_h_omni.vision.patch_size
16
16
nemotron_h_omni.vision.projector.scale_factor
2
2
nemotron_h_omni.vision.use_thumbnail
true
true
tokenizer.ggml.add_bos_token
false
false
tokenizer.ggml.add_eos_token
false
false
tokenizer.ggml.add_unknown_token
false
false
tokenizer.ggml.bos_token_id
1
1
tokenizer.ggml.eos_token_id
11
11
tokenizer.ggml.eos_token_ids
[2, 11]
[2, 11]
tokenizer.ggml.merges
[Ġ Ġ, Ġ t, e r, i n, Ġ ĠĠĠ, ...]
[Ġ Ġ, Ġ t, e r, i n, Ġ ĠĠĠ, ...]
tokenizer.ggml.model
gpt2
gpt2
tokenizer.ggml.pre
default
default
tokenizer.ggml.scores
[0, 1, 2, 3, 4, ...]
[0, 1, 2, 3, 4, ...]
tokenizer.ggml.token_type
[3, 3, 3, 3, 3, ...]
[3, 3, 3, 3, 3, ...]
tokenizer.ggml.tokens
[<unk>, <s>, </s>, [INST], [/INST], ...]
[<unk>, <s>, </s>, [INST], [/INST], ...]
tokenizer.ggml.unknown_token_id
0
0
Tensor
Name
Type
Shape
token_embd.weight
Q8_0
Q8_0
[2688, 131072]
a.blk.0
a.blk.0.attn_bias_u
BF16
BF16
[128, 8]
a.blk.0.attn_bias_v
BF16
BF16
[128, 8]
a.blk.0.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.0.attn_norm.bias
F32
F32
[1024]
a.blk.0.attn_norm.weight
F32
F32
[1024]
a.blk.0.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.0.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.0.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.0.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.0.conv_bn.bias
F32
F32
[1024]
a.blk.0.conv_bn.running_mean
F32
F32
[1024]
a.blk.0.conv_bn.running_var
F32
F32
[1024]
a.blk.0.conv_bn.weight
F32
F32
[1024]
a.blk.0.conv_dw.weight
F32
F32
[9, 1024]
a.blk.0.conv_norm.bias
F32
F32
[1024]
a.blk.0.conv_norm.weight
F32
F32
[1024]
a.blk.0.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.0.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.0.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.0.ffn1_norm.bias
F32
F32
[1024]
a.blk.0.ffn1_norm.weight
F32
F32
[1024]
a.blk.0.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.0.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.0.ffn2_norm.bias
F32
F32
[1024]
a.blk.0.ffn2_norm.weight
F32
F32
[1024]
a.blk.0.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.0.out_norm.bias
F32
F32
[1024]
a.blk.0.out_norm.weight
F32
F32
[1024]
a.blk.1
a.blk.1.attn_bias_u
BF16
BF16
[128, 8]
a.blk.1.attn_bias_v
BF16
BF16
[128, 8]
a.blk.1.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.1.attn_norm.bias
F32
F32
[1024]
a.blk.1.attn_norm.weight
F32
F32
[1024]
a.blk.1.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.1.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.1.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.1.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.1.conv_bn.bias
F32
F32
[1024]
a.blk.1.conv_bn.running_mean
F32
F32
[1024]
a.blk.1.conv_bn.running_var
F32
F32
[1024]
a.blk.1.conv_bn.weight
F32
F32
[1024]
a.blk.1.conv_dw.weight
F32
F32
[9, 1024]
a.blk.1.conv_norm.bias
F32
F32
[1024]
a.blk.1.conv_norm.weight
F32
F32
[1024]
a.blk.1.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.1.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.1.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.1.ffn1_norm.bias
F32
F32
[1024]
a.blk.1.ffn1_norm.weight
F32
F32
[1024]
a.blk.1.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.1.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.1.ffn2_norm.bias
F32
F32
[1024]
a.blk.1.ffn2_norm.weight
F32
F32
[1024]
a.blk.1.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.1.out_norm.bias
F32
F32
[1024]
a.blk.1.out_norm.weight
F32
F32
[1024]
a.blk.2
a.blk.2.attn_bias_u
BF16
BF16
[128, 8]
a.blk.2.attn_bias_v
BF16
BF16
[128, 8]
a.blk.2.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.2.attn_norm.bias
F32
F32
[1024]
a.blk.2.attn_norm.weight
F32
F32
[1024]
a.blk.2.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.2.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.2.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.2.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.2.conv_bn.bias
F32
F32
[1024]
a.blk.2.conv_bn.running_mean
F32
F32
[1024]
a.blk.2.conv_bn.running_var
F32
F32
[1024]
a.blk.2.conv_bn.weight
F32
F32
[1024]
a.blk.2.conv_dw.weight
F32
F32
[9, 1024]
a.blk.2.conv_norm.bias
F32
F32
[1024]
a.blk.2.conv_norm.weight
F32
F32
[1024]
a.blk.2.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.2.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.2.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.2.ffn1_norm.bias
F32
F32
[1024]
a.blk.2.ffn1_norm.weight
F32
F32
[1024]
a.blk.2.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.2.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.2.ffn2_norm.bias
F32
F32
[1024]
a.blk.2.ffn2_norm.weight
F32
F32
[1024]
a.blk.2.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.2.out_norm.bias
F32
F32
[1024]
a.blk.2.out_norm.weight
F32
F32
[1024]
a.blk.3
a.blk.3.attn_bias_u
BF16
BF16
[128, 8]
a.blk.3.attn_bias_v
BF16
BF16
[128, 8]
a.blk.3.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.3.attn_norm.bias
F32
F32
[1024]
a.blk.3.attn_norm.weight
F32
F32
[1024]
a.blk.3.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.3.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.3.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.3.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.3.conv_bn.bias
F32
F32
[1024]
a.blk.3.conv_bn.running_mean
F32
F32
[1024]
a.blk.3.conv_bn.running_var
F32
F32
[1024]
a.blk.3.conv_bn.weight
F32
F32
[1024]
a.blk.3.conv_dw.weight
F32
F32
[9, 1024]
a.blk.3.conv_norm.bias
F32
F32
[1024]
a.blk.3.conv_norm.weight
F32
F32
[1024]
a.blk.3.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.3.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.3.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.3.ffn1_norm.bias
F32
F32
[1024]
a.blk.3.ffn1_norm.weight
F32
F32
[1024]
a.blk.3.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.3.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.3.ffn2_norm.bias
F32
F32
[1024]
a.blk.3.ffn2_norm.weight
F32
F32
[1024]
a.blk.3.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.3.out_norm.bias
F32
F32
[1024]
a.blk.3.out_norm.weight
F32
F32
[1024]
a.blk.4
a.blk.4.attn_bias_u
BF16
BF16
[128, 8]
a.blk.4.attn_bias_v
BF16
BF16
[128, 8]
a.blk.4.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.4.attn_norm.bias
F32
F32
[1024]
a.blk.4.attn_norm.weight
F32
F32
[1024]
a.blk.4.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.4.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.4.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.4.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.4.conv_bn.bias
F32
F32
[1024]
a.blk.4.conv_bn.running_mean
F32
F32
[1024]
a.blk.4.conv_bn.running_var
F32
F32
[1024]
a.blk.4.conv_bn.weight
F32
F32
[1024]
a.blk.4.conv_dw.weight
F32
F32
[9, 1024]
a.blk.4.conv_norm.bias
F32
F32
[1024]
a.blk.4.conv_norm.weight
F32
F32
[1024]
a.blk.4.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.4.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.4.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.4.ffn1_norm.bias
F32
F32
[1024]
a.blk.4.ffn1_norm.weight
F32
F32
[1024]
a.blk.4.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.4.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.4.ffn2_norm.bias
F32
F32
[1024]
a.blk.4.ffn2_norm.weight
F32
F32
[1024]
a.blk.4.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.4.out_norm.bias
F32
F32
[1024]
a.blk.4.out_norm.weight
F32
F32
[1024]
a.blk.5
a.blk.5.attn_bias_u
BF16
BF16
[128, 8]
a.blk.5.attn_bias_v
BF16
BF16
[128, 8]
a.blk.5.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.5.attn_norm.bias
F32
F32
[1024]
a.blk.5.attn_norm.weight
F32
F32
[1024]
a.blk.5.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.5.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.5.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.5.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.5.conv_bn.bias
F32
F32
[1024]
a.blk.5.conv_bn.running_mean
F32
F32
[1024]
a.blk.5.conv_bn.running_var
F32
F32
[1024]
a.blk.5.conv_bn.weight
F32
F32
[1024]
a.blk.5.conv_dw.weight
F32
F32
[9, 1024]
a.blk.5.conv_norm.bias
F32
F32
[1024]
a.blk.5.conv_norm.weight
F32
F32
[1024]
a.blk.5.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.5.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.5.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.5.ffn1_norm.bias
F32
F32
[1024]
a.blk.5.ffn1_norm.weight
F32
F32
[1024]
a.blk.5.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.5.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.5.ffn2_norm.bias
F32
F32
[1024]
a.blk.5.ffn2_norm.weight
F32
F32
[1024]
a.blk.5.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.5.out_norm.bias
F32
F32
[1024]
a.blk.5.out_norm.weight
F32
F32
[1024]
a.blk.6
a.blk.6.attn_bias_u
BF16
BF16
[128, 8]
a.blk.6.attn_bias_v
BF16
BF16
[128, 8]
a.blk.6.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.6.attn_norm.bias
F32
F32
[1024]
a.blk.6.attn_norm.weight
F32
F32
[1024]
a.blk.6.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.6.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.6.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.6.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.6.conv_bn.bias
F32
F32
[1024]
a.blk.6.conv_bn.running_mean
F32
F32
[1024]
a.blk.6.conv_bn.running_var
F32
F32
[1024]
a.blk.6.conv_bn.weight
F32
F32
[1024]
a.blk.6.conv_dw.weight
F32
F32
[9, 1024]
a.blk.6.conv_norm.bias
F32
F32
[1024]
a.blk.6.conv_norm.weight
F32
F32
[1024]
a.blk.6.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.6.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.6.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.6.ffn1_norm.bias
F32
F32
[1024]
a.blk.6.ffn1_norm.weight
F32
F32
[1024]
a.blk.6.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.6.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.6.ffn2_norm.bias
F32
F32
[1024]
a.blk.6.ffn2_norm.weight
F32
F32
[1024]
a.blk.6.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.6.out_norm.bias
F32
F32
[1024]
a.blk.6.out_norm.weight
F32
F32
[1024]
a.blk.7
a.blk.7.attn_bias_u
BF16
BF16
[128, 8]
a.blk.7.attn_bias_v
BF16
BF16
[128, 8]
a.blk.7.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.7.attn_norm.bias
F32
F32
[1024]
a.blk.7.attn_norm.weight
F32
F32
[1024]
a.blk.7.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.7.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.7.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.7.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.7.conv_bn.bias
F32
F32
[1024]
a.blk.7.conv_bn.running_mean
F32
F32
[1024]
a.blk.7.conv_bn.running_var
F32
F32
[1024]
a.blk.7.conv_bn.weight
F32
F32
[1024]
a.blk.7.conv_dw.weight
F32
F32
[9, 1024]
a.blk.7.conv_norm.bias
F32
F32
[1024]
a.blk.7.conv_norm.weight
F32
F32
[1024]
a.blk.7.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.7.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.7.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.7.ffn1_norm.bias
F32
F32
[1024]
a.blk.7.ffn1_norm.weight
F32
F32
[1024]
a.blk.7.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.7.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.7.ffn2_norm.bias
F32
F32
[1024]
a.blk.7.ffn2_norm.weight
F32
F32
[1024]
a.blk.7.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.7.out_norm.bias
F32
F32
[1024]
a.blk.7.out_norm.weight
F32
F32
[1024]
a.blk.8
a.blk.8.attn_bias_u
BF16
BF16
[128, 8]
a.blk.8.attn_bias_v
BF16
BF16
[128, 8]
a.blk.8.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.8.attn_norm.bias
F32
F32
[1024]
a.blk.8.attn_norm.weight
F32
F32
[1024]
a.blk.8.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.8.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.8.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.8.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.8.conv_bn.bias
F32
F32
[1024]
a.blk.8.conv_bn.running_mean
F32
F32
[1024]
a.blk.8.conv_bn.running_var
F32
F32
[1024]
a.blk.8.conv_bn.weight
F32
F32
[1024]
a.blk.8.conv_dw.weight
F32
F32
[9, 1024]
a.blk.8.conv_norm.bias
F32
F32
[1024]
a.blk.8.conv_norm.weight
F32
F32
[1024]
a.blk.8.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.8.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.8.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.8.ffn1_norm.bias
F32
F32
[1024]
a.blk.8.ffn1_norm.weight
F32
F32
[1024]
a.blk.8.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.8.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.8.ffn2_norm.bias
F32
F32
[1024]
a.blk.8.ffn2_norm.weight
F32
F32
[1024]
a.blk.8.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.8.out_norm.bias
F32
F32
[1024]
a.blk.8.out_norm.weight
F32
F32
[1024]
a.blk.9
a.blk.9.attn_bias_u
BF16
BF16
[128, 8]
a.blk.9.attn_bias_v
BF16
BF16
[128, 8]
a.blk.9.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.9.attn_norm.bias
F32
F32
[1024]
a.blk.9.attn_norm.weight
F32
F32
[1024]
a.blk.9.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.9.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.9.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.9.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.9.conv_bn.bias
F32
F32
[1024]
a.blk.9.conv_bn.running_mean
F32
F32
[1024]
a.blk.9.conv_bn.running_var
F32
F32
[1024]
a.blk.9.conv_bn.weight
F32
F32
[1024]
a.blk.9.conv_dw.weight
F32
F32
[9, 1024]
a.blk.9.conv_norm.bias
F32
F32
[1024]
a.blk.9.conv_norm.weight
F32
F32
[1024]
a.blk.9.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.9.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.9.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.9.ffn1_norm.bias
F32
F32
[1024]
a.blk.9.ffn1_norm.weight
F32
F32
[1024]
a.blk.9.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.9.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.9.ffn2_norm.bias
F32
F32
[1024]
a.blk.9.ffn2_norm.weight
F32
F32
[1024]
a.blk.9.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.9.out_norm.bias
F32
F32
[1024]
a.blk.9.out_norm.weight
F32
F32
[1024]
a.blk.10
a.blk.10.attn_bias_u
BF16
BF16
[128, 8]
a.blk.10.attn_bias_v
BF16
BF16
[128, 8]
a.blk.10.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.10.attn_norm.bias
F32
F32
[1024]
a.blk.10.attn_norm.weight
F32
F32
[1024]
a.blk.10.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.10.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.10.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.10.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.10.conv_bn.bias
F32
F32
[1024]
a.blk.10.conv_bn.running_mean
F32
F32
[1024]
a.blk.10.conv_bn.running_var
F32
F32
[1024]
a.blk.10.conv_bn.weight
F32
F32
[1024]
a.blk.10.conv_dw.weight
F32
F32
[9, 1024]
a.blk.10.conv_norm.bias
F32
F32
[1024]
a.blk.10.conv_norm.weight
F32
F32
[1024]
a.blk.10.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.10.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.10.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.10.ffn1_norm.bias
F32
F32
[1024]
a.blk.10.ffn1_norm.weight
F32
F32
[1024]
a.blk.10.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.10.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.10.ffn2_norm.bias
F32
F32
[1024]
a.blk.10.ffn2_norm.weight
F32
F32
[1024]
a.blk.10.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.10.out_norm.bias
F32
F32
[1024]
a.blk.10.out_norm.weight
F32
F32
[1024]
a.blk.11
a.blk.11.attn_bias_u
BF16
BF16
[128, 8]
a.blk.11.attn_bias_v
BF16
BF16
[128, 8]
a.blk.11.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.11.attn_norm.bias
F32
F32
[1024]
a.blk.11.attn_norm.weight
F32
F32
[1024]
a.blk.11.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.11.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.11.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.11.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.11.conv_bn.bias
F32
F32
[1024]
a.blk.11.conv_bn.running_mean
F32
F32
[1024]
a.blk.11.conv_bn.running_var
F32
F32
[1024]
a.blk.11.conv_bn.weight
F32
F32
[1024]
a.blk.11.conv_dw.weight
F32
F32
[9, 1024]
a.blk.11.conv_norm.bias
F32
F32
[1024]
a.blk.11.conv_norm.weight
F32
F32
[1024]
a.blk.11.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.11.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.11.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.11.ffn1_norm.bias
F32
F32
[1024]
a.blk.11.ffn1_norm.weight
F32
F32
[1024]
a.blk.11.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.11.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.11.ffn2_norm.bias
F32
F32
[1024]
a.blk.11.ffn2_norm.weight
F32
F32
[1024]
a.blk.11.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.11.out_norm.bias
F32
F32
[1024]
a.blk.11.out_norm.weight
F32
F32
[1024]
a.blk.12
a.blk.12.attn_bias_u
BF16
BF16
[128, 8]
a.blk.12.attn_bias_v
BF16
BF16
[128, 8]
a.blk.12.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.12.attn_norm.bias
F32
F32
[1024]
a.blk.12.attn_norm.weight
F32
F32
[1024]
a.blk.12.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.12.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.12.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.12.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.12.conv_bn.bias
F32
F32
[1024]
a.blk.12.conv_bn.running_mean
F32
F32
[1024]
a.blk.12.conv_bn.running_var
F32
F32
[1024]
a.blk.12.conv_bn.weight
F32
F32
[1024]
a.blk.12.conv_dw.weight
F32
F32
[9, 1024]
a.blk.12.conv_norm.bias
F32
F32
[1024]
a.blk.12.conv_norm.weight
F32
F32
[1024]
a.blk.12.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.12.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.12.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.12.ffn1_norm.bias
F32
F32
[1024]
a.blk.12.ffn1_norm.weight
F32
F32
[1024]
a.blk.12.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.12.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.12.ffn2_norm.bias
F32
F32
[1024]
a.blk.12.ffn2_norm.weight
F32
F32
[1024]
a.blk.12.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.12.out_norm.bias
F32
F32
[1024]
a.blk.12.out_norm.weight
F32
F32
[1024]
a.blk.13
a.blk.13.attn_bias_u
BF16
BF16
[128, 8]
a.blk.13.attn_bias_v
BF16
BF16
[128, 8]
a.blk.13.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.13.attn_norm.bias
F32
F32
[1024]
a.blk.13.attn_norm.weight
F32
F32
[1024]
a.blk.13.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.13.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.13.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.13.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.13.conv_bn.bias
F32
F32
[1024]
a.blk.13.conv_bn.running_mean
F32
F32
[1024]
a.blk.13.conv_bn.running_var
F32
F32
[1024]
a.blk.13.conv_bn.weight
F32
F32
[1024]
a.blk.13.conv_dw.weight
F32
F32
[9, 1024]
a.blk.13.conv_norm.bias
F32
F32
[1024]
a.blk.13.conv_norm.weight
F32
F32
[1024]
a.blk.13.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.13.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.13.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.13.ffn1_norm.bias
F32
F32
[1024]
a.blk.13.ffn1_norm.weight
F32
F32
[1024]
a.blk.13.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.13.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.13.ffn2_norm.bias
F32
F32
[1024]
a.blk.13.ffn2_norm.weight
F32
F32
[1024]
a.blk.13.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.13.out_norm.bias
F32
F32
[1024]
a.blk.13.out_norm.weight
F32
F32
[1024]
a.blk.14
a.blk.14.attn_bias_u
BF16
BF16
[128, 8]
a.blk.14.attn_bias_v
BF16
BF16
[128, 8]
a.blk.14.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.14.attn_norm.bias
F32
F32
[1024]
a.blk.14.attn_norm.weight
F32
F32
[1024]
a.blk.14.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.14.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.14.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.14.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.14.conv_bn.bias
F32
F32
[1024]
a.blk.14.conv_bn.running_mean
F32
F32
[1024]
a.blk.14.conv_bn.running_var
F32
F32
[1024]
a.blk.14.conv_bn.weight
F32
F32
[1024]
a.blk.14.conv_dw.weight
F32
F32
[9, 1024]
a.blk.14.conv_norm.bias
F32
F32
[1024]
a.blk.14.conv_norm.weight
F32
F32
[1024]
a.blk.14.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.14.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.14.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.14.ffn1_norm.bias
F32
F32
[1024]
a.blk.14.ffn1_norm.weight
F32
F32
[1024]
a.blk.14.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.14.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.14.ffn2_norm.bias
F32
F32
[1024]
a.blk.14.ffn2_norm.weight
F32
F32
[1024]
a.blk.14.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.14.out_norm.bias
F32
F32
[1024]
a.blk.14.out_norm.weight
F32
F32
[1024]
a.blk.15
a.blk.15.attn_bias_u
BF16
BF16
[128, 8]
a.blk.15.attn_bias_v
BF16
BF16
[128, 8]
a.blk.15.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.15.attn_norm.bias
F32
F32
[1024]
a.blk.15.attn_norm.weight
F32
F32
[1024]
a.blk.15.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.15.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.15.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.15.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.15.conv_bn.bias
F32
F32
[1024]
a.blk.15.conv_bn.running_mean
F32
F32
[1024]
a.blk.15.conv_bn.running_var
F32
F32
[1024]
a.blk.15.conv_bn.weight
F32
F32
[1024]
a.blk.15.conv_dw.weight
F32
F32
[9, 1024]
a.blk.15.conv_norm.bias
F32
F32
[1024]
a.blk.15.conv_norm.weight
F32
F32
[1024]
a.blk.15.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.15.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.15.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.15.ffn1_norm.bias
F32
F32
[1024]
a.blk.15.ffn1_norm.weight
F32
F32
[1024]
a.blk.15.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.15.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.15.ffn2_norm.bias
F32
F32
[1024]
a.blk.15.ffn2_norm.weight
F32
F32
[1024]
a.blk.15.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.15.out_norm.bias
F32
F32
[1024]
a.blk.15.out_norm.weight
F32
F32
[1024]
a.blk.16
a.blk.16.attn_bias_u
BF16
BF16
[128, 8]
a.blk.16.attn_bias_v
BF16
BF16
[128, 8]
a.blk.16.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.16.attn_norm.bias
F32
F32
[1024]
a.blk.16.attn_norm.weight
F32
F32
[1024]
a.blk.16.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.16.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.16.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.16.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.16.conv_bn.bias
F32
F32
[1024]
a.blk.16.conv_bn.running_mean
F32
F32
[1024]
a.blk.16.conv_bn.running_var
F32
F32
[1024]
a.blk.16.conv_bn.weight
F32
F32
[1024]
a.blk.16.conv_dw.weight
F32
F32
[9, 1024]
a.blk.16.conv_norm.bias
F32
F32
[1024]
a.blk.16.conv_norm.weight
F32
F32
[1024]
a.blk.16.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.16.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.16.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.16.ffn1_norm.bias
F32
F32
[1024]
a.blk.16.ffn1_norm.weight
F32
F32
[1024]
a.blk.16.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.16.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.16.ffn2_norm.bias
F32
F32
[1024]
a.blk.16.ffn2_norm.weight
F32
F32
[1024]
a.blk.16.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.16.out_norm.bias
F32
F32
[1024]
a.blk.16.out_norm.weight
F32
F32
[1024]
a.blk.17
a.blk.17.attn_bias_u
BF16
BF16
[128, 8]
a.blk.17.attn_bias_v
BF16
BF16
[128, 8]
a.blk.17.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.17.attn_norm.bias
F32
F32
[1024]
a.blk.17.attn_norm.weight
F32
F32
[1024]
a.blk.17.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.17.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.17.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.17.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.17.conv_bn.bias
F32
F32
[1024]
a.blk.17.conv_bn.running_mean
F32
F32
[1024]
a.blk.17.conv_bn.running_var
F32
F32
[1024]
a.blk.17.conv_bn.weight
F32
F32
[1024]
a.blk.17.conv_dw.weight
F32
F32
[9, 1024]
a.blk.17.conv_norm.bias
F32
F32
[1024]
a.blk.17.conv_norm.weight
F32
F32
[1024]
a.blk.17.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.17.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.17.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.17.ffn1_norm.bias
F32
F32
[1024]
a.blk.17.ffn1_norm.weight
F32
F32
[1024]
a.blk.17.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.17.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.17.ffn2_norm.bias
F32
F32
[1024]
a.blk.17.ffn2_norm.weight
F32
F32
[1024]
a.blk.17.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.17.out_norm.bias
F32
F32
[1024]
a.blk.17.out_norm.weight
F32
F32
[1024]
a.blk.18
a.blk.18.attn_bias_u
BF16
BF16
[128, 8]
a.blk.18.attn_bias_v
BF16
BF16
[128, 8]
a.blk.18.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.18.attn_norm.bias
F32
F32
[1024]
a.blk.18.attn_norm.weight
F32
F32
[1024]
a.blk.18.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.18.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.18.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.18.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.18.conv_bn.bias
F32
F32
[1024]
a.blk.18.conv_bn.running_mean
F32
F32
[1024]
a.blk.18.conv_bn.running_var
F32
F32
[1024]
a.blk.18.conv_bn.weight
F32
F32
[1024]
a.blk.18.conv_dw.weight
F32
F32
[9, 1024]
a.blk.18.conv_norm.bias
F32
F32
[1024]
a.blk.18.conv_norm.weight
F32
F32
[1024]
a.blk.18.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.18.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.18.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.18.ffn1_norm.bias
F32
F32
[1024]
a.blk.18.ffn1_norm.weight
F32
F32
[1024]
a.blk.18.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.18.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.18.ffn2_norm.bias
F32
F32
[1024]
a.blk.18.ffn2_norm.weight
F32
F32
[1024]
a.blk.18.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.18.out_norm.bias
F32
F32
[1024]
a.blk.18.out_norm.weight
F32
F32
[1024]
a.blk.19
a.blk.19.attn_bias_u
BF16
BF16
[128, 8]
a.blk.19.attn_bias_v
BF16
BF16
[128, 8]
a.blk.19.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.19.attn_norm.bias
F32
F32
[1024]
a.blk.19.attn_norm.weight
F32
F32
[1024]
a.blk.19.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.19.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.19.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.19.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.19.conv_bn.bias
F32
F32
[1024]
a.blk.19.conv_bn.running_mean
F32
F32
[1024]
a.blk.19.conv_bn.running_var
F32
F32
[1024]
a.blk.19.conv_bn.weight
F32
F32
[1024]
a.blk.19.conv_dw.weight
F32
F32
[9, 1024]
a.blk.19.conv_norm.bias
F32
F32
[1024]
a.blk.19.conv_norm.weight
F32
F32
[1024]
a.blk.19.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.19.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.19.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.19.ffn1_norm.bias
F32
F32
[1024]
a.blk.19.ffn1_norm.weight
F32
F32
[1024]
a.blk.19.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.19.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.19.ffn2_norm.bias
F32
F32
[1024]
a.blk.19.ffn2_norm.weight
F32
F32
[1024]
a.blk.19.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.19.out_norm.bias
F32
F32
[1024]
a.blk.19.out_norm.weight
F32
F32
[1024]
a.blk.20
a.blk.20.attn_bias_u
BF16
BF16
[128, 8]
a.blk.20.attn_bias_v
BF16
BF16
[128, 8]
a.blk.20.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.20.attn_norm.bias
F32
F32
[1024]
a.blk.20.attn_norm.weight
F32
F32
[1024]
a.blk.20.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.20.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.20.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.20.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.20.conv_bn.bias
F32
F32
[1024]
a.blk.20.conv_bn.running_mean
F32
F32
[1024]
a.blk.20.conv_bn.running_var
F32
F32
[1024]
a.blk.20.conv_bn.weight
F32
F32
[1024]
a.blk.20.conv_dw.weight
F32
F32
[9, 1024]
a.blk.20.conv_norm.bias
F32
F32
[1024]
a.blk.20.conv_norm.weight
F32
F32
[1024]
a.blk.20.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.20.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.20.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.20.ffn1_norm.bias
F32
F32
[1024]
a.blk.20.ffn1_norm.weight
F32
F32
[1024]
a.blk.20.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.20.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.20.ffn2_norm.bias
F32
F32
[1024]
a.blk.20.ffn2_norm.weight
F32
F32
[1024]
a.blk.20.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.20.out_norm.bias
F32
F32
[1024]
a.blk.20.out_norm.weight
F32
F32
[1024]
a.blk.21
a.blk.21.attn_bias_u
BF16
BF16
[128, 8]
a.blk.21.attn_bias_v
BF16
BF16
[128, 8]
a.blk.21.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.21.attn_norm.bias
F32
F32
[1024]
a.blk.21.attn_norm.weight
F32
F32
[1024]
a.blk.21.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.21.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.21.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.21.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.21.conv_bn.bias
F32
F32
[1024]
a.blk.21.conv_bn.running_mean
F32
F32
[1024]
a.blk.21.conv_bn.running_var
F32
F32
[1024]
a.blk.21.conv_bn.weight
F32
F32
[1024]
a.blk.21.conv_dw.weight
F32
F32
[9, 1024]
a.blk.21.conv_norm.bias
F32
F32
[1024]
a.blk.21.conv_norm.weight
F32
F32
[1024]
a.blk.21.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.21.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.21.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.21.ffn1_norm.bias
F32
F32
[1024]
a.blk.21.ffn1_norm.weight
F32
F32
[1024]
a.blk.21.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.21.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.21.ffn2_norm.bias
F32
F32
[1024]
a.blk.21.ffn2_norm.weight
F32
F32
[1024]
a.blk.21.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.21.out_norm.bias
F32
F32
[1024]
a.blk.21.out_norm.weight
F32
F32
[1024]
a.blk.22
a.blk.22.attn_bias_u
BF16
BF16
[128, 8]
a.blk.22.attn_bias_v
BF16
BF16
[128, 8]
a.blk.22.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.22.attn_norm.bias
F32
F32
[1024]
a.blk.22.attn_norm.weight
F32
F32
[1024]
a.blk.22.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.22.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.22.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.22.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.22.conv_bn.bias
F32
F32
[1024]
a.blk.22.conv_bn.running_mean
F32
F32
[1024]
a.blk.22.conv_bn.running_var
F32
F32
[1024]
a.blk.22.conv_bn.weight
F32
F32
[1024]
a.blk.22.conv_dw.weight
F32
F32
[9, 1024]
a.blk.22.conv_norm.bias
F32
F32
[1024]
a.blk.22.conv_norm.weight
F32
F32
[1024]
a.blk.22.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.22.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.22.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.22.ffn1_norm.bias
F32
F32
[1024]
a.blk.22.ffn1_norm.weight
F32
F32
[1024]
a.blk.22.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.22.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.22.ffn2_norm.bias
F32
F32
[1024]
a.blk.22.ffn2_norm.weight
F32
F32
[1024]
a.blk.22.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.22.out_norm.bias
F32
F32
[1024]
a.blk.22.out_norm.weight
F32
F32
[1024]
a.blk.23
a.blk.23.attn_bias_u
BF16
BF16
[128, 8]
a.blk.23.attn_bias_v
BF16
BF16
[128, 8]
a.blk.23.attn_k.weight
BF16
BF16
[1024, 1024]
a.blk.23.attn_norm.bias
F32
F32
[1024]
a.blk.23.attn_norm.weight
F32
F32
[1024]
a.blk.23.attn_out.weight
BF16
BF16
[1024, 1024]
a.blk.23.attn_q.weight
BF16
BF16
[1024, 1024]
a.blk.23.attn_rel_k.weight
BF16
BF16
[1024, 1024]
a.blk.23.attn_v.weight
BF16
BF16
[1024, 1024]
a.blk.23.conv_bn.bias
F32
F32
[1024]
a.blk.23.conv_bn.running_mean
F32
F32
[1024]
a.blk.23.conv_bn.running_var
F32
F32
[1024]
a.blk.23.conv_bn.weight
F32
F32
[1024]
a.blk.23.conv_dw.weight
F32
F32
[9, 1024]
a.blk.23.conv_norm.bias
F32
F32
[1024]
a.blk.23.conv_norm.weight
F32
F32
[1024]
a.blk.23.conv_pw1.weight
BF16
BF16
[1024, 2048]
a.blk.23.conv_pw2.weight
BF16
BF16
[1024, 1024]
a.blk.23.ffn1_down.weight
BF16
BF16
[4096, 1024]
a.blk.23.ffn1_norm.bias
F32
F32
[1024]
a.blk.23.ffn1_norm.weight
F32
F32
[1024]
a.blk.23.ffn1_up.weight
BF16
BF16
[1024, 4096]
a.blk.23.ffn2_down.weight
BF16
BF16
[4096, 1024]
a.blk.23.ffn2_norm.bias
F32
F32
[1024]
a.blk.23.ffn2_norm.weight
F32
F32
[1024]
a.blk.23.ffn2_up.weight
BF16
BF16
[1024, 4096]
a.blk.23.out_norm.bias
F32
F32
[1024]
a.blk.23.out_norm.weight
F32
F32
[1024]
a.feature_extractor.fb
F32
F32
[257, 128, 1]
a.feature_extractor.window
F32
F32
[400]
a.subsampling.conv0.bias
F32
F32
[256]
a.subsampling.conv0.weight
F32
F32
[3, 3, 1, 256]
a.subsampling.dw1.bias
F32
F32
[256]
a.subsampling.dw1.weight
F32
F32
[3, 3, 1, 256]
a.subsampling.dw2.bias
F32
F32
[256]
a.subsampling.dw2.weight
F32
F32
[3, 3, 1, 256]
a.subsampling.linear.bias
F32
F32
[1024]
a.subsampling.linear.weight
F32
F32
[4096, 1024]
a.subsampling.pw1.bias
F32
F32
[256]
a.subsampling.pw1.weight
F32
F32
[1, 1, 256, 256]
a.subsampling.pw2.bias
F32
F32
[256]
a.subsampling.pw2.weight
F32
F32
[1, 1, 256, 256]
blk.0
blk.0.attn_norm.weight
F32
F32
[2688]
blk.0.ssm_a
F32
F32
[1, 64]
blk.0.ssm_conv1d.bias
F32
F32
[6144]
blk.0.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.0.ssm_d
F32
F32
[1, 64]
blk.0.ssm_dt.bias
F32
F32
[64]
blk.0.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.0.ssm_norm.weight
F32
F32
[512, 8]
blk.0.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.1
blk.1.attn_norm.weight
F32
F32
[2688]
blk.1.exp_probs_b.bias
F32
F32
[128]
blk.1.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.1.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.1.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.1.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.1.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.2
blk.2.attn_norm.weight
F32
F32
[2688]
blk.2.ssm_a
F32
F32
[1, 64]
blk.2.ssm_conv1d.bias
F32
F32
[6144]
blk.2.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.2.ssm_d
F32
F32
[1, 64]
blk.2.ssm_dt.bias
F32
F32
[64]
blk.2.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.2.ssm_norm.weight
F32
F32
[512, 8]
blk.2.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.3
blk.3.attn_norm.weight
F32
F32
[2688]
blk.3.exp_probs_b.bias
F32
F32
[128]
blk.3.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.3.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.3.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.3.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.3.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.4
blk.4.attn_norm.weight
F32
F32
[2688]
blk.4.ssm_a
F32
F32
[1, 64]
blk.4.ssm_conv1d.bias
F32
F32
[6144]
blk.4.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.4.ssm_d
F32
F32
[1, 64]
blk.4.ssm_dt.bias
F32
F32
[64]
blk.4.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.4.ssm_norm.weight
F32
F32
[512, 8]
blk.4.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.5
blk.5.attn_k.weight
Q8_0
Q8_0
[2688, 256]
blk.5.attn_norm.weight
F32
F32
[2688]
blk.5.attn_output.weight
Q8_0
Q8_0
[4096, 2688]
blk.5.attn_q.weight
Q8_0
Q8_0
[2688, 4096]
blk.5.attn_v.weight
Q8_0
Q8_0
[2688, 256]
blk.6
blk.6.attn_norm.weight
F32
F32
[2688]
blk.6.exp_probs_b.bias
F32
F32
[128]
blk.6.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.6.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.6.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.6.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.6.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.7
blk.7.attn_norm.weight
F32
F32
[2688]
blk.7.ssm_a
F32
F32
[1, 64]
blk.7.ssm_conv1d.bias
F32
F32
[6144]
blk.7.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.7.ssm_d
F32
F32
[1, 64]
blk.7.ssm_dt.bias
F32
F32
[64]
blk.7.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.7.ssm_norm.weight
F32
F32
[512, 8]
blk.7.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.8
blk.8.attn_norm.weight
F32
F32
[2688]
blk.8.exp_probs_b.bias
F32
F32
[128]
blk.8.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.8.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.8.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.8.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.8.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.9
blk.9.attn_norm.weight
F32
F32
[2688]
blk.9.ssm_a
F32
F32
[1, 64]
blk.9.ssm_conv1d.bias
F32
F32
[6144]
blk.9.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.9.ssm_d
F32
F32
[1, 64]
blk.9.ssm_dt.bias
F32
F32
[64]
blk.9.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.9.ssm_norm.weight
F32
F32
[512, 8]
blk.9.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.10
blk.10.attn_norm.weight
F32
F32
[2688]
blk.10.exp_probs_b.bias
F32
F32
[128]
blk.10.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.10.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.10.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.10.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.10.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.11
blk.11.attn_norm.weight
F32
F32
[2688]
blk.11.ssm_a
F32
F32
[1, 64]
blk.11.ssm_conv1d.bias
F32
F32
[6144]
blk.11.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.11.ssm_d
F32
F32
[1, 64]
blk.11.ssm_dt.bias
F32
F32
[64]
blk.11.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.11.ssm_norm.weight
F32
F32
[512, 8]
blk.11.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.12
blk.12.attn_k.weight
Q8_0
Q8_0
[2688, 256]
blk.12.attn_norm.weight
F32
F32
[2688]
blk.12.attn_output.weight
Q8_0
Q8_0
[4096, 2688]
blk.12.attn_q.weight
Q8_0
Q8_0
[2688, 4096]
blk.12.attn_v.weight
Q8_0
Q8_0
[2688, 256]
blk.13
blk.13.attn_norm.weight
F32
F32
[2688]
blk.13.exp_probs_b.bias
F32
F32
[128]
blk.13.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.13.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.13.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.13.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.13.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.14
blk.14.attn_norm.weight
F32
F32
[2688]
blk.14.ssm_a
F32
F32
[1, 64]
blk.14.ssm_conv1d.bias
F32
F32
[6144]
blk.14.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.14.ssm_d
F32
F32
[1, 64]
blk.14.ssm_dt.bias
F32
F32
[64]
blk.14.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.14.ssm_norm.weight
F32
F32
[512, 8]
blk.14.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.15
blk.15.attn_norm.weight
F32
F32
[2688]
blk.15.exp_probs_b.bias
F32
F32
[128]
blk.15.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.15.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.15.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.15.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.15.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.16
blk.16.attn_norm.weight
F32
F32
[2688]
blk.16.ssm_a
F32
F32
[1, 64]
blk.16.ssm_conv1d.bias
F32
F32
[6144]
blk.16.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.16.ssm_d
F32
F32
[1, 64]
blk.16.ssm_dt.bias
F32
F32
[64]
blk.16.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.16.ssm_norm.weight
F32
F32
[512, 8]
blk.16.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.17
blk.17.attn_norm.weight
F32
F32
[2688]
blk.17.exp_probs_b.bias
F32
F32
[128]
blk.17.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.17.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.17.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.17.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.17.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.18
blk.18.attn_norm.weight
F32
F32
[2688]
blk.18.ssm_a
F32
F32
[1, 64]
blk.18.ssm_conv1d.bias
F32
F32
[6144]
blk.18.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.18.ssm_d
F32
F32
[1, 64]
blk.18.ssm_dt.bias
F32
F32
[64]
blk.18.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.18.ssm_norm.weight
F32
F32
[512, 8]
blk.18.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.19
blk.19.attn_k.weight
Q8_0
Q8_0
[2688, 256]
blk.19.attn_norm.weight
F32
F32
[2688]
blk.19.attn_output.weight
Q8_0
Q8_0
[4096, 2688]
blk.19.attn_q.weight
Q8_0
Q8_0
[2688, 4096]
blk.19.attn_v.weight
Q8_0
Q8_0
[2688, 256]
blk.20
blk.20.attn_norm.weight
F32
F32
[2688]
blk.20.exp_probs_b.bias
F32
F32
[128]
blk.20.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.20.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.20.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.20.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.20.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.21
blk.21.attn_norm.weight
F32
F32
[2688]
blk.21.ssm_a
F32
F32
[1, 64]
blk.21.ssm_conv1d.bias
F32
F32
[6144]
blk.21.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.21.ssm_d
F32
F32
[1, 64]
blk.21.ssm_dt.bias
F32
F32
[64]
blk.21.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.21.ssm_norm.weight
F32
F32
[512, 8]
blk.21.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.22
blk.22.attn_norm.weight
F32
F32
[2688]
blk.22.exp_probs_b.bias
F32
F32
[128]
blk.22.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.22.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.22.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.22.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.22.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.23
blk.23.attn_norm.weight
F32
F32
[2688]
blk.23.ssm_a
F32
F32
[1, 64]
blk.23.ssm_conv1d.bias
F32
F32
[6144]
blk.23.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.23.ssm_d
F32
F32
[1, 64]
blk.23.ssm_dt.bias
F32
F32
[64]
blk.23.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.23.ssm_norm.weight
F32
F32
[512, 8]
blk.23.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.24
blk.24.attn_norm.weight
F32
F32
[2688]
blk.24.exp_probs_b.bias
F32
F32
[128]
blk.24.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.24.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.24.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.24.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.24.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.25
blk.25.attn_norm.weight
F32
F32
[2688]
blk.25.ssm_a
F32
F32
[1, 64]
blk.25.ssm_conv1d.bias
F32
F32
[6144]
blk.25.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.25.ssm_d
F32
F32
[1, 64]
blk.25.ssm_dt.bias
F32
F32
[64]
blk.25.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.25.ssm_norm.weight
F32
F32
[512, 8]
blk.25.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.26
blk.26.attn_k.weight
Q8_0
Q8_0
[2688, 256]
blk.26.attn_norm.weight
F32
F32
[2688]
blk.26.attn_output.weight
Q8_0
Q8_0
[4096, 2688]
blk.26.attn_q.weight
Q8_0
Q8_0
[2688, 4096]
blk.26.attn_v.weight
Q8_0
Q8_0
[2688, 256]
blk.27
blk.27.attn_norm.weight
F32
F32
[2688]
blk.27.exp_probs_b.bias
F32
F32
[128]
blk.27.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.27.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.27.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.27.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.27.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.28
blk.28.attn_norm.weight
F32
F32
[2688]
blk.28.ssm_a
F32
F32
[1, 64]
blk.28.ssm_conv1d.bias
F32
F32
[6144]
blk.28.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.28.ssm_d
F32
F32
[1, 64]
blk.28.ssm_dt.bias
F32
F32
[64]
blk.28.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.28.ssm_norm.weight
F32
F32
[512, 8]
blk.28.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.29
blk.29.attn_norm.weight
F32
F32
[2688]
blk.29.exp_probs_b.bias
F32
F32
[128]
blk.29.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.29.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.29.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.29.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.29.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.30
blk.30.attn_norm.weight
F32
F32
[2688]
blk.30.ssm_a
F32
F32
[1, 64]
blk.30.ssm_conv1d.bias
F32
F32
[6144]
blk.30.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.30.ssm_d
F32
F32
[1, 64]
blk.30.ssm_dt.bias
F32
F32
[64]
blk.30.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.30.ssm_norm.weight
F32
F32
[512, 8]
blk.30.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.31
blk.31.attn_norm.weight
F32
F32
[2688]
blk.31.exp_probs_b.bias
F32
F32
[128]
blk.31.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.31.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.31.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.31.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.31.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.32
blk.32.attn_norm.weight
F32
F32
[2688]
blk.32.ssm_a
F32
F32
[1, 64]
blk.32.ssm_conv1d.bias
F32
F32
[6144]
blk.32.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.32.ssm_d
F32
F32
[1, 64]
blk.32.ssm_dt.bias
F32
F32
[64]
blk.32.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.32.ssm_norm.weight
F32
F32
[512, 8]
blk.32.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.33
blk.33.attn_k.weight
Q8_0
Q8_0
[2688, 256]
blk.33.attn_norm.weight
F32
F32
[2688]
blk.33.attn_output.weight
Q8_0
Q8_0
[4096, 2688]
blk.33.attn_q.weight
Q8_0
Q8_0
[2688, 4096]
blk.33.attn_v.weight
Q8_0
Q8_0
[2688, 256]
blk.34
blk.34.attn_norm.weight
F32
F32
[2688]
blk.34.exp_probs_b.bias
F32
F32
[128]
blk.34.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.34.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.34.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.34.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.34.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.35
blk.35.attn_norm.weight
F32
F32
[2688]
blk.35.ssm_a
F32
F32
[1, 64]
blk.35.ssm_conv1d.bias
F32
F32
[6144]
blk.35.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.35.ssm_d
F32
F32
[1, 64]
blk.35.ssm_dt.bias
F32
F32
[64]
blk.35.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.35.ssm_norm.weight
F32
F32
[512, 8]
blk.35.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.36
blk.36.attn_norm.weight
F32
F32
[2688]
blk.36.exp_probs_b.bias
F32
F32
[128]
blk.36.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.36.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.36.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.36.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.36.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.37
blk.37.attn_norm.weight
F32
F32
[2688]
blk.37.ssm_a
F32
F32
[1, 64]
blk.37.ssm_conv1d.bias
F32
F32
[6144]
blk.37.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.37.ssm_d
F32
F32
[1, 64]
blk.37.ssm_dt.bias
F32
F32
[64]
blk.37.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.37.ssm_norm.weight
F32
F32
[512, 8]
blk.37.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.38
blk.38.attn_norm.weight
F32
F32
[2688]
blk.38.exp_probs_b.bias
F32
F32
[128]
blk.38.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.38.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.38.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.38.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.38.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.39
blk.39.attn_norm.weight
F32
F32
[2688]
blk.39.ssm_a
F32
F32
[1, 64]
blk.39.ssm_conv1d.bias
F32
F32
[6144]
blk.39.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.39.ssm_d
F32
F32
[1, 64]
blk.39.ssm_dt.bias
F32
F32
[64]
blk.39.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.39.ssm_norm.weight
F32
F32
[512, 8]
blk.39.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.40
blk.40.attn_norm.weight
F32
F32
[2688]
blk.40.exp_probs_b.bias
F32
F32
[128]
blk.40.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.40.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.40.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.40.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.40.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.41
blk.41.attn_norm.weight
F32
F32
[2688]
blk.41.ssm_a
F32
F32
[1, 64]
blk.41.ssm_conv1d.bias
F32
F32
[6144]
blk.41.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.41.ssm_d
F32
F32
[1, 64]
blk.41.ssm_dt.bias
F32
F32
[64]
blk.41.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.41.ssm_norm.weight
F32
F32
[512, 8]
blk.41.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.42
blk.42.attn_k.weight
Q8_0
Q8_0
[2688, 256]
blk.42.attn_norm.weight
F32
F32
[2688]
blk.42.attn_output.weight
Q8_0
Q8_0
[4096, 2688]
blk.42.attn_q.weight
Q8_0
Q8_0
[2688, 4096]
blk.42.attn_v.weight
Q8_0
Q8_0
[2688, 256]
blk.43
blk.43.attn_norm.weight
F32
F32
[2688]
blk.43.exp_probs_b.bias
F32
F32
[128]
blk.43.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.43.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.43.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.43.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.43.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.44
blk.44.attn_norm.weight
F32
F32
[2688]
blk.44.ssm_a
F32
F32
[1, 64]
blk.44.ssm_conv1d.bias
F32
F32
[6144]
blk.44.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.44.ssm_d
F32
F32
[1, 64]
blk.44.ssm_dt.bias
F32
F32
[64]
blk.44.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.44.ssm_norm.weight
F32
F32
[512, 8]
blk.44.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.45
blk.45.attn_norm.weight
F32
F32
[2688]
blk.45.exp_probs_b.bias
F32
F32
[128]
blk.45.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.45.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.45.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.45.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.45.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.46
blk.46.attn_norm.weight
F32
F32
[2688]
blk.46.ssm_a
F32
F32
[1, 64]
blk.46.ssm_conv1d.bias
F32
F32
[6144]
blk.46.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.46.ssm_d
F32
F32
[1, 64]
blk.46.ssm_dt.bias
F32
F32
[64]
blk.46.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.46.ssm_norm.weight
F32
F32
[512, 8]
blk.46.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.47
blk.47.attn_norm.weight
F32
F32
[2688]
blk.47.exp_probs_b.bias
F32
F32
[128]
blk.47.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.47.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.47.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.47.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.47.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.48
blk.48.attn_norm.weight
F32
F32
[2688]
blk.48.ssm_a
F32
F32
[1, 64]
blk.48.ssm_conv1d.bias
F32
F32
[6144]
blk.48.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.48.ssm_d
F32
F32
[1, 64]
blk.48.ssm_dt.bias
F32
F32
[64]
blk.48.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.48.ssm_norm.weight
F32
F32
[512, 8]
blk.48.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.49
blk.49.attn_norm.weight
F32
F32
[2688]
blk.49.exp_probs_b.bias
F32
F32
[128]
blk.49.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.49.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.49.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.49.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.49.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
blk.50
blk.50.attn_norm.weight
F32
F32
[2688]
blk.50.ssm_a
F32
F32
[1, 64]
blk.50.ssm_conv1d.bias
F32
F32
[6144]
blk.50.ssm_conv1d.weight
F32
F32
[4, 6144]
blk.50.ssm_d
F32
F32
[1, 64]
blk.50.ssm_dt.bias
F32
F32
[64]
blk.50.ssm_in.weight
Q8_0
Q8_0
[2688, 10304]
blk.50.ssm_norm.weight
F32
F32
[512, 8]
blk.50.ssm_out.weight
Q8_0
Q8_0
[4096, 2688]
blk.51
blk.51.attn_norm.weight
F32
F32
[2688]
blk.51.exp_probs_b.bias
F32
F32
[128]
blk.51.ffn_down_exps.weight
Q8_0
Q8_0
[1856, 2688, 128]
blk.51.ffn_down_shexp.weight
Q8_0
Q8_0
[3712, 2688]
blk.51.ffn_gate_inp.weight
F32
F32
[2688, 128]
blk.51.ffn_up_exps.weight
Q8_0
Q8_0
[2688, 1856, 128]
blk.51.ffn_up_shexp.weight
Q8_0
Q8_0
[2688, 3712]
mm.1.weight
F16
F16
[5120, 20480]
mm.2.weight
F16
F16
[20480, 2688]
mm.a.1.weight
F16
F16
[1024, 4096]
mm.a.2.weight
F16
F16
[4096, 2688]
mm.a.norm.weight
F32
F32
[1024]
mm.norm.weight
F32
F32
[5120]
output.weight
Q8_0
Q8_0
[2688, 131072]
v.blk.0
v.blk.0.attn_k.bias
F32
F32
[1280]
v.blk.0.attn_k.weight
F16
F16
[1280, 1280]
v.blk.0.attn_out.bias
F32
F32
[1280]
v.blk.0.attn_out.weight
F16
F16
[1280, 1280]
v.blk.0.attn_q.bias
F32
F32
[1280]
v.blk.0.attn_q.weight
F16
F16
[1280, 1280]
v.blk.0.attn_v.bias
F32
F32
[1280]
v.blk.0.attn_v.weight
F16
F16
[1280, 1280]
v.blk.0.ffn_down.bias
F32
F32
[1280]
v.blk.0.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.0.ffn_up.bias
F32
F32
[5120]
v.blk.0.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.0.ln1.bias
F32
F32
[1280]
v.blk.0.ln1.weight
F32
F32
[1280]
v.blk.0.ln2.bias
F32
F32
[1280]
v.blk.0.ln2.weight
F32
F32
[1280]
v.blk.1
v.blk.1.attn_k.bias
F32
F32
[1280]
v.blk.1.attn_k.weight
F16
F16
[1280, 1280]
v.blk.1.attn_out.bias
F32
F32
[1280]
v.blk.1.attn_out.weight
F16
F16
[1280, 1280]
v.blk.1.attn_q.bias
F32
F32
[1280]
v.blk.1.attn_q.weight
F16
F16
[1280, 1280]
v.blk.1.attn_v.bias
F32
F32
[1280]
v.blk.1.attn_v.weight
F16
F16
[1280, 1280]
v.blk.1.ffn_down.bias
F32
F32
[1280]
v.blk.1.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.1.ffn_up.bias
F32
F32
[5120]
v.blk.1.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.1.ln1.bias
F32
F32
[1280]
v.blk.1.ln1.weight
F32
F32
[1280]
v.blk.1.ln2.bias
F32
F32
[1280]
v.blk.1.ln2.weight
F32
F32
[1280]
v.blk.2
v.blk.2.attn_k.bias
F32
F32
[1280]
v.blk.2.attn_k.weight
F16
F16
[1280, 1280]
v.blk.2.attn_out.bias
F32
F32
[1280]
v.blk.2.attn_out.weight
F16
F16
[1280, 1280]
v.blk.2.attn_q.bias
F32
F32
[1280]
v.blk.2.attn_q.weight
F16
F16
[1280, 1280]
v.blk.2.attn_v.bias
F32
F32
[1280]
v.blk.2.attn_v.weight
F16
F16
[1280, 1280]
v.blk.2.ffn_down.bias
F32
F32
[1280]
v.blk.2.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.2.ffn_up.bias
F32
F32
[5120]
v.blk.2.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.2.ln1.bias
F32
F32
[1280]
v.blk.2.ln1.weight
F32
F32
[1280]
v.blk.2.ln2.bias
F32
F32
[1280]
v.blk.2.ln2.weight
F32
F32
[1280]
v.blk.3
v.blk.3.attn_k.bias
F32
F32
[1280]
v.blk.3.attn_k.weight
F16
F16
[1280, 1280]
v.blk.3.attn_out.bias
F32
F32
[1280]
v.blk.3.attn_out.weight
F16
F16
[1280, 1280]
v.blk.3.attn_q.bias
F32
F32
[1280]
v.blk.3.attn_q.weight
F16
F16
[1280, 1280]
v.blk.3.attn_v.bias
F32
F32
[1280]
v.blk.3.attn_v.weight
F16
F16
[1280, 1280]
v.blk.3.ffn_down.bias
F32
F32
[1280]
v.blk.3.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.3.ffn_up.bias
F32
F32
[5120]
v.blk.3.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.3.ln1.bias
F32
F32
[1280]
v.blk.3.ln1.weight
F32
F32
[1280]
v.blk.3.ln2.bias
F32
F32
[1280]
v.blk.3.ln2.weight
F32
F32
[1280]
v.blk.4
v.blk.4.attn_k.bias
F32
F32
[1280]
v.blk.4.attn_k.weight
F16
F16
[1280, 1280]
v.blk.4.attn_out.bias
F32
F32
[1280]
v.blk.4.attn_out.weight
F16
F16
[1280, 1280]
v.blk.4.attn_q.bias
F32
F32
[1280]
v.blk.4.attn_q.weight
F16
F16
[1280, 1280]
v.blk.4.attn_v.bias
F32
F32
[1280]
v.blk.4.attn_v.weight
F16
F16
[1280, 1280]
v.blk.4.ffn_down.bias
F32
F32
[1280]
v.blk.4.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.4.ffn_up.bias
F32
F32
[5120]
v.blk.4.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.4.ln1.bias
F32
F32
[1280]
v.blk.4.ln1.weight
F32
F32
[1280]
v.blk.4.ln2.bias
F32
F32
[1280]
v.blk.4.ln2.weight
F32
F32
[1280]
v.blk.5
v.blk.5.attn_k.bias
F32
F32
[1280]
v.blk.5.attn_k.weight
F16
F16
[1280, 1280]
v.blk.5.attn_out.bias
F32
F32
[1280]
v.blk.5.attn_out.weight
F16
F16
[1280, 1280]
v.blk.5.attn_q.bias
F32
F32
[1280]
v.blk.5.attn_q.weight
F16
F16
[1280, 1280]
v.blk.5.attn_v.bias
F32
F32
[1280]
v.blk.5.attn_v.weight
F16
F16
[1280, 1280]
v.blk.5.ffn_down.bias
F32
F32
[1280]
v.blk.5.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.5.ffn_up.bias
F32
F32
[5120]
v.blk.5.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.5.ln1.bias
F32
F32
[1280]
v.blk.5.ln1.weight
F32
F32
[1280]
v.blk.5.ln2.bias
F32
F32
[1280]
v.blk.5.ln2.weight
F32
F32
[1280]
v.blk.6
v.blk.6.attn_k.bias
F32
F32
[1280]
v.blk.6.attn_k.weight
F16
F16
[1280, 1280]
v.blk.6.attn_out.bias
F32
F32
[1280]
v.blk.6.attn_out.weight
F16
F16
[1280, 1280]
v.blk.6.attn_q.bias
F32
F32
[1280]
v.blk.6.attn_q.weight
F16
F16
[1280, 1280]
v.blk.6.attn_v.bias
F32
F32
[1280]
v.blk.6.attn_v.weight
F16
F16
[1280, 1280]
v.blk.6.ffn_down.bias
F32
F32
[1280]
v.blk.6.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.6.ffn_up.bias
F32
F32
[5120]
v.blk.6.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.6.ln1.bias
F32
F32
[1280]
v.blk.6.ln1.weight
F32
F32
[1280]
v.blk.6.ln2.bias
F32
F32
[1280]
v.blk.6.ln2.weight
F32
F32
[1280]
v.blk.7
v.blk.7.attn_k.bias
F32
F32
[1280]
v.blk.7.attn_k.weight
F16
F16
[1280, 1280]
v.blk.7.attn_out.bias
F32
F32
[1280]
v.blk.7.attn_out.weight
F16
F16
[1280, 1280]
v.blk.7.attn_q.bias
F32
F32
[1280]
v.blk.7.attn_q.weight
F16
F16
[1280, 1280]
v.blk.7.attn_v.bias
F32
F32
[1280]
v.blk.7.attn_v.weight
F16
F16
[1280, 1280]
v.blk.7.ffn_down.bias
F32
F32
[1280]
v.blk.7.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.7.ffn_up.bias
F32
F32
[5120]
v.blk.7.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.7.ln1.bias
F32
F32
[1280]
v.blk.7.ln1.weight
F32
F32
[1280]
v.blk.7.ln2.bias
F32
F32
[1280]
v.blk.7.ln2.weight
F32
F32
[1280]
v.blk.8
v.blk.8.attn_k.bias
F32
F32
[1280]
v.blk.8.attn_k.weight
F16
F16
[1280, 1280]
v.blk.8.attn_out.bias
F32
F32
[1280]
v.blk.8.attn_out.weight
F16
F16
[1280, 1280]
v.blk.8.attn_q.bias
F32
F32
[1280]
v.blk.8.attn_q.weight
F16
F16
[1280, 1280]
v.blk.8.attn_v.bias
F32
F32
[1280]
v.blk.8.attn_v.weight
F16
F16
[1280, 1280]
v.blk.8.ffn_down.bias
F32
F32
[1280]
v.blk.8.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.8.ffn_up.bias
F32
F32
[5120]
v.blk.8.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.8.ln1.bias
F32
F32
[1280]
v.blk.8.ln1.weight
F32
F32
[1280]
v.blk.8.ln2.bias
F32
F32
[1280]
v.blk.8.ln2.weight
F32
F32
[1280]
v.blk.9
v.blk.9.attn_k.bias
F32
F32
[1280]
v.blk.9.attn_k.weight
F16
F16
[1280, 1280]
v.blk.9.attn_out.bias
F32
F32
[1280]
v.blk.9.attn_out.weight
F16
F16
[1280, 1280]
v.blk.9.attn_q.bias
F32
F32
[1280]
v.blk.9.attn_q.weight
F16
F16
[1280, 1280]
v.blk.9.attn_v.bias
F32
F32
[1280]
v.blk.9.attn_v.weight
F16
F16
[1280, 1280]
v.blk.9.ffn_down.bias
F32
F32
[1280]
v.blk.9.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.9.ffn_up.bias
F32
F32
[5120]
v.blk.9.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.9.ln1.bias
F32
F32
[1280]
v.blk.9.ln1.weight
F32
F32
[1280]
v.blk.9.ln2.bias
F32
F32
[1280]
v.blk.9.ln2.weight
F32
F32
[1280]
v.blk.10
v.blk.10.attn_k.bias
F32
F32
[1280]
v.blk.10.attn_k.weight
F16
F16
[1280, 1280]
v.blk.10.attn_out.bias
F32
F32
[1280]
v.blk.10.attn_out.weight
F16
F16
[1280, 1280]
v.blk.10.attn_q.bias
F32
F32
[1280]
v.blk.10.attn_q.weight
F16
F16
[1280, 1280]
v.blk.10.attn_v.bias
F32
F32
[1280]
v.blk.10.attn_v.weight
F16
F16
[1280, 1280]
v.blk.10.ffn_down.bias
F32
F32
[1280]
v.blk.10.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.10.ffn_up.bias
F32
F32
[5120]
v.blk.10.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.10.ln1.bias
F32
F32
[1280]
v.blk.10.ln1.weight
F32
F32
[1280]
v.blk.10.ln2.bias
F32
F32
[1280]
v.blk.10.ln2.weight
F32
F32
[1280]
v.blk.11
v.blk.11.attn_k.bias
F32
F32
[1280]
v.blk.11.attn_k.weight
F16
F16
[1280, 1280]
v.blk.11.attn_out.bias
F32
F32
[1280]
v.blk.11.attn_out.weight
F16
F16
[1280, 1280]
v.blk.11.attn_q.bias
F32
F32
[1280]
v.blk.11.attn_q.weight
F16
F16
[1280, 1280]
v.blk.11.attn_v.bias
F32
F32
[1280]
v.blk.11.attn_v.weight
F16
F16
[1280, 1280]
v.blk.11.ffn_down.bias
F32
F32
[1280]
v.blk.11.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.11.ffn_up.bias
F32
F32
[5120]
v.blk.11.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.11.ln1.bias
F32
F32
[1280]
v.blk.11.ln1.weight
F32
F32
[1280]
v.blk.11.ln2.bias
F32
F32
[1280]
v.blk.11.ln2.weight
F32
F32
[1280]
v.blk.12
v.blk.12.attn_k.bias
F32
F32
[1280]
v.blk.12.attn_k.weight
F16
F16
[1280, 1280]
v.blk.12.attn_out.bias
F32
F32
[1280]
v.blk.12.attn_out.weight
F16
F16
[1280, 1280]
v.blk.12.attn_q.bias
F32
F32
[1280]
v.blk.12.attn_q.weight
F16
F16
[1280, 1280]
v.blk.12.attn_v.bias
F32
F32
[1280]
v.blk.12.attn_v.weight
F16
F16
[1280, 1280]
v.blk.12.ffn_down.bias
F32
F32
[1280]
v.blk.12.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.12.ffn_up.bias
F32
F32
[5120]
v.blk.12.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.12.ln1.bias
F32
F32
[1280]
v.blk.12.ln1.weight
F32
F32
[1280]
v.blk.12.ln2.bias
F32
F32
[1280]
v.blk.12.ln2.weight
F32
F32
[1280]
v.blk.13
v.blk.13.attn_k.bias
F32
F32
[1280]
v.blk.13.attn_k.weight
F16
F16
[1280, 1280]
v.blk.13.attn_out.bias
F32
F32
[1280]
v.blk.13.attn_out.weight
F16
F16
[1280, 1280]
v.blk.13.attn_q.bias
F32
F32
[1280]
v.blk.13.attn_q.weight
F16
F16
[1280, 1280]
v.blk.13.attn_v.bias
F32
F32
[1280]
v.blk.13.attn_v.weight
F16
F16
[1280, 1280]
v.blk.13.ffn_down.bias
F32
F32
[1280]
v.blk.13.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.13.ffn_up.bias
F32
F32
[5120]
v.blk.13.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.13.ln1.bias
F32
F32
[1280]
v.blk.13.ln1.weight
F32
F32
[1280]
v.blk.13.ln2.bias
F32
F32
[1280]
v.blk.13.ln2.weight
F32
F32
[1280]
v.blk.14
v.blk.14.attn_k.bias
F32
F32
[1280]
v.blk.14.attn_k.weight
F16
F16
[1280, 1280]
v.blk.14.attn_out.bias
F32
F32
[1280]
v.blk.14.attn_out.weight
F16
F16
[1280, 1280]
v.blk.14.attn_q.bias
F32
F32
[1280]
v.blk.14.attn_q.weight
F16
F16
[1280, 1280]
v.blk.14.attn_v.bias
F32
F32
[1280]
v.blk.14.attn_v.weight
F16
F16
[1280, 1280]
v.blk.14.ffn_down.bias
F32
F32
[1280]
v.blk.14.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.14.ffn_up.bias
F32
F32
[5120]
v.blk.14.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.14.ln1.bias
F32
F32
[1280]
v.blk.14.ln1.weight
F32
F32
[1280]
v.blk.14.ln2.bias
F32
F32
[1280]
v.blk.14.ln2.weight
F32
F32
[1280]
v.blk.15
v.blk.15.attn_k.bias
F32
F32
[1280]
v.blk.15.attn_k.weight
F16
F16
[1280, 1280]
v.blk.15.attn_out.bias
F32
F32
[1280]
v.blk.15.attn_out.weight
F16
F16
[1280, 1280]
v.blk.15.attn_q.bias
F32
F32
[1280]
v.blk.15.attn_q.weight
F16
F16
[1280, 1280]
v.blk.15.attn_v.bias
F32
F32
[1280]
v.blk.15.attn_v.weight
F16
F16
[1280, 1280]
v.blk.15.ffn_down.bias
F32
F32
[1280]
v.blk.15.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.15.ffn_up.bias
F32
F32
[5120]
v.blk.15.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.15.ln1.bias
F32
F32
[1280]
v.blk.15.ln1.weight
F32
F32
[1280]
v.blk.15.ln2.bias
F32
F32
[1280]
v.blk.15.ln2.weight
F32
F32
[1280]
v.blk.16
v.blk.16.attn_k.bias
F32
F32
[1280]
v.blk.16.attn_k.weight
F16
F16
[1280, 1280]
v.blk.16.attn_out.bias
F32
F32
[1280]
v.blk.16.attn_out.weight
F16
F16
[1280, 1280]
v.blk.16.attn_q.bias
F32
F32
[1280]
v.blk.16.attn_q.weight
F16
F16
[1280, 1280]
v.blk.16.attn_v.bias
F32
F32
[1280]
v.blk.16.attn_v.weight
F16
F16
[1280, 1280]
v.blk.16.ffn_down.bias
F32
F32
[1280]
v.blk.16.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.16.ffn_up.bias
F32
F32
[5120]
v.blk.16.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.16.ln1.bias
F32
F32
[1280]
v.blk.16.ln1.weight
F32
F32
[1280]
v.blk.16.ln2.bias
F32
F32
[1280]
v.blk.16.ln2.weight
F32
F32
[1280]
v.blk.17
v.blk.17.attn_k.bias
F32
F32
[1280]
v.blk.17.attn_k.weight
F16
F16
[1280, 1280]
v.blk.17.attn_out.bias
F32
F32
[1280]
v.blk.17.attn_out.weight
F16
F16
[1280, 1280]
v.blk.17.attn_q.bias
F32
F32
[1280]
v.blk.17.attn_q.weight
F16
F16
[1280, 1280]
v.blk.17.attn_v.bias
F32
F32
[1280]
v.blk.17.attn_v.weight
F16
F16
[1280, 1280]
v.blk.17.ffn_down.bias
F32
F32
[1280]
v.blk.17.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.17.ffn_up.bias
F32
F32
[5120]
v.blk.17.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.17.ln1.bias
F32
F32
[1280]
v.blk.17.ln1.weight
F32
F32
[1280]
v.blk.17.ln2.bias
F32
F32
[1280]
v.blk.17.ln2.weight
F32
F32
[1280]
v.blk.18
v.blk.18.attn_k.bias
F32
F32
[1280]
v.blk.18.attn_k.weight
F16
F16
[1280, 1280]
v.blk.18.attn_out.bias
F32
F32
[1280]
v.blk.18.attn_out.weight
F16
F16
[1280, 1280]
v.blk.18.attn_q.bias
F32
F32
[1280]
v.blk.18.attn_q.weight
F16
F16
[1280, 1280]
v.blk.18.attn_v.bias
F32
F32
[1280]
v.blk.18.attn_v.weight
F16
F16
[1280, 1280]
v.blk.18.ffn_down.bias
F32
F32
[1280]
v.blk.18.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.18.ffn_up.bias
F32
F32
[5120]
v.blk.18.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.18.ln1.bias
F32
F32
[1280]
v.blk.18.ln1.weight
F32
F32
[1280]
v.blk.18.ln2.bias
F32
F32
[1280]
v.blk.18.ln2.weight
F32
F32
[1280]
v.blk.19
v.blk.19.attn_k.bias
F32
F32
[1280]
v.blk.19.attn_k.weight
F16
F16
[1280, 1280]
v.blk.19.attn_out.bias
F32
F32
[1280]
v.blk.19.attn_out.weight
F16
F16
[1280, 1280]
v.blk.19.attn_q.bias
F32
F32
[1280]
v.blk.19.attn_q.weight
F16
F16
[1280, 1280]
v.blk.19.attn_v.bias
F32
F32
[1280]
v.blk.19.attn_v.weight
F16
F16
[1280, 1280]
v.blk.19.ffn_down.bias
F32
F32
[1280]
v.blk.19.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.19.ffn_up.bias
F32
F32
[5120]
v.blk.19.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.19.ln1.bias
F32
F32
[1280]
v.blk.19.ln1.weight
F32
F32
[1280]
v.blk.19.ln2.bias
F32
F32
[1280]
v.blk.19.ln2.weight
F32
F32
[1280]
v.blk.20
v.blk.20.attn_k.bias
F32
F32
[1280]
v.blk.20.attn_k.weight
F16
F16
[1280, 1280]
v.blk.20.attn_out.bias
F32
F32
[1280]
v.blk.20.attn_out.weight
F16
F16
[1280, 1280]
v.blk.20.attn_q.bias
F32
F32
[1280]
v.blk.20.attn_q.weight
F16
F16
[1280, 1280]
v.blk.20.attn_v.bias
F32
F32
[1280]
v.blk.20.attn_v.weight
F16
F16
[1280, 1280]
v.blk.20.ffn_down.bias
F32
F32
[1280]
v.blk.20.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.20.ffn_up.bias
F32
F32
[5120]
v.blk.20.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.20.ln1.bias
F32
F32
[1280]
v.blk.20.ln1.weight
F32
F32
[1280]
v.blk.20.ln2.bias
F32
F32
[1280]
v.blk.20.ln2.weight
F32
F32
[1280]
v.blk.21
v.blk.21.attn_k.bias
F32
F32
[1280]
v.blk.21.attn_k.weight
F16
F16
[1280, 1280]
v.blk.21.attn_out.bias
F32
F32
[1280]
v.blk.21.attn_out.weight
F16
F16
[1280, 1280]
v.blk.21.attn_q.bias
F32
F32
[1280]
v.blk.21.attn_q.weight
F16
F16
[1280, 1280]
v.blk.21.attn_v.bias
F32
F32
[1280]
v.blk.21.attn_v.weight
F16
F16
[1280, 1280]
v.blk.21.ffn_down.bias
F32
F32
[1280]
v.blk.21.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.21.ffn_up.bias
F32
F32
[5120]
v.blk.21.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.21.ln1.bias
F32
F32
[1280]
v.blk.21.ln1.weight
F32
F32
[1280]
v.blk.21.ln2.bias
F32
F32
[1280]
v.blk.21.ln2.weight
F32
F32
[1280]
v.blk.22
v.blk.22.attn_k.bias
F32
F32
[1280]
v.blk.22.attn_k.weight
F16
F16
[1280, 1280]
v.blk.22.attn_out.bias
F32
F32
[1280]
v.blk.22.attn_out.weight
F16
F16
[1280, 1280]
v.blk.22.attn_q.bias
F32
F32
[1280]
v.blk.22.attn_q.weight
F16
F16
[1280, 1280]
v.blk.22.attn_v.bias
F32
F32
[1280]
v.blk.22.attn_v.weight
F16
F16
[1280, 1280]
v.blk.22.ffn_down.bias
F32
F32
[1280]
v.blk.22.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.22.ffn_up.bias
F32
F32
[5120]
v.blk.22.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.22.ln1.bias
F32
F32
[1280]
v.blk.22.ln1.weight
F32
F32
[1280]
v.blk.22.ln2.bias
F32
F32
[1280]
v.blk.22.ln2.weight
F32
F32
[1280]
v.blk.23
v.blk.23.attn_k.bias
F32
F32
[1280]
v.blk.23.attn_k.weight
F16
F16
[1280, 1280]
v.blk.23.attn_out.bias
F32
F32
[1280]
v.blk.23.attn_out.weight
F16
F16
[1280, 1280]
v.blk.23.attn_q.bias
F32
F32
[1280]
v.blk.23.attn_q.weight
F16
F16
[1280, 1280]
v.blk.23.attn_v.bias
F32
F32
[1280]
v.blk.23.attn_v.weight
F16
F16
[1280, 1280]
v.blk.23.ffn_down.bias
F32
F32
[1280]
v.blk.23.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.23.ffn_up.bias
F32
F32
[5120]
v.blk.23.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.23.ln1.bias
F32
F32
[1280]
v.blk.23.ln1.weight
F32
F32
[1280]
v.blk.23.ln2.bias
F32
F32
[1280]
v.blk.23.ln2.weight
F32
F32
[1280]
v.blk.24
v.blk.24.attn_k.bias
F32
F32
[1280]
v.blk.24.attn_k.weight
F16
F16
[1280, 1280]
v.blk.24.attn_out.bias
F32
F32
[1280]
v.blk.24.attn_out.weight
F16
F16
[1280, 1280]
v.blk.24.attn_q.bias
F32
F32
[1280]
v.blk.24.attn_q.weight
F16
F16
[1280, 1280]
v.blk.24.attn_v.bias
F32
F32
[1280]
v.blk.24.attn_v.weight
F16
F16
[1280, 1280]
v.blk.24.ffn_down.bias
F32
F32
[1280]
v.blk.24.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.24.ffn_up.bias
F32
F32
[5120]
v.blk.24.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.24.ln1.bias
F32
F32
[1280]
v.blk.24.ln1.weight
F32
F32
[1280]
v.blk.24.ln2.bias
F32
F32
[1280]
v.blk.24.ln2.weight
F32
F32
[1280]
v.blk.25
v.blk.25.attn_k.bias
F32
F32
[1280]
v.blk.25.attn_k.weight
F16
F16
[1280, 1280]
v.blk.25.attn_out.bias
F32
F32
[1280]
v.blk.25.attn_out.weight
F16
F16
[1280, 1280]
v.blk.25.attn_q.bias
F32
F32
[1280]
v.blk.25.attn_q.weight
F16
F16
[1280, 1280]
v.blk.25.attn_v.bias
F32
F32
[1280]
v.blk.25.attn_v.weight
F16
F16
[1280, 1280]
v.blk.25.ffn_down.bias
F32
F32
[1280]
v.blk.25.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.25.ffn_up.bias
F32
F32
[5120]
v.blk.25.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.25.ln1.bias
F32
F32
[1280]
v.blk.25.ln1.weight
F32
F32
[1280]
v.blk.25.ln2.bias
F32
F32
[1280]
v.blk.25.ln2.weight
F32
F32
[1280]
v.blk.26
v.blk.26.attn_k.bias
F32
F32
[1280]
v.blk.26.attn_k.weight
F16
F16
[1280, 1280]
v.blk.26.attn_out.bias
F32
F32
[1280]
v.blk.26.attn_out.weight
F16
F16
[1280, 1280]
v.blk.26.attn_q.bias
F32
F32
[1280]
v.blk.26.attn_q.weight
F16
F16
[1280, 1280]
v.blk.26.attn_v.bias
F32
F32
[1280]
v.blk.26.attn_v.weight
F16
F16
[1280, 1280]
v.blk.26.ffn_down.bias
F32
F32
[1280]
v.blk.26.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.26.ffn_up.bias
F32
F32
[5120]
v.blk.26.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.26.ln1.bias
F32
F32
[1280]
v.blk.26.ln1.weight
F32
F32
[1280]
v.blk.26.ln2.bias
F32
F32
[1280]
v.blk.26.ln2.weight
F32
F32
[1280]
v.blk.27
v.blk.27.attn_k.bias
F32
F32
[1280]
v.blk.27.attn_k.weight
F16
F16
[1280, 1280]
v.blk.27.attn_out.bias
F32
F32
[1280]
v.blk.27.attn_out.weight
F16
F16
[1280, 1280]
v.blk.27.attn_q.bias
F32
F32
[1280]
v.blk.27.attn_q.weight
F16
F16
[1280, 1280]
v.blk.27.attn_v.bias
F32
F32
[1280]
v.blk.27.attn_v.weight
F16
F16
[1280, 1280]
v.blk.27.ffn_down.bias
F32
F32
[1280]
v.blk.27.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.27.ffn_up.bias
F32
F32
[5120]
v.blk.27.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.27.ln1.bias
F32
F32
[1280]
v.blk.27.ln1.weight
F32
F32
[1280]
v.blk.27.ln2.bias
F32
F32
[1280]
v.blk.27.ln2.weight
F32
F32
[1280]
v.blk.28
v.blk.28.attn_k.bias
F32
F32
[1280]
v.blk.28.attn_k.weight
F16
F16
[1280, 1280]
v.blk.28.attn_out.bias
F32
F32
[1280]
v.blk.28.attn_out.weight
F16
F16
[1280, 1280]
v.blk.28.attn_q.bias
F32
F32
[1280]
v.blk.28.attn_q.weight
F16
F16
[1280, 1280]
v.blk.28.attn_v.bias
F32
F32
[1280]
v.blk.28.attn_v.weight
F16
F16
[1280, 1280]
v.blk.28.ffn_down.bias
F32
F32
[1280]
v.blk.28.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.28.ffn_up.bias
F32
F32
[5120]
v.blk.28.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.28.ln1.bias
F32
F32
[1280]
v.blk.28.ln1.weight
F32
F32
[1280]
v.blk.28.ln2.bias
F32
F32
[1280]
v.blk.28.ln2.weight
F32
F32
[1280]
v.blk.29
v.blk.29.attn_k.bias
F32
F32
[1280]
v.blk.29.attn_k.weight
F16
F16
[1280, 1280]
v.blk.29.attn_out.bias
F32
F32
[1280]
v.blk.29.attn_out.weight
F16
F16
[1280, 1280]
v.blk.29.attn_q.bias
F32
F32
[1280]
v.blk.29.attn_q.weight
F16
F16
[1280, 1280]
v.blk.29.attn_v.bias
F32
F32
[1280]
v.blk.29.attn_v.weight
F16
F16
[1280, 1280]
v.blk.29.ffn_down.bias
F32
F32
[1280]
v.blk.29.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.29.ffn_up.bias
F32
F32
[5120]
v.blk.29.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.29.ln1.bias
F32
F32
[1280]
v.blk.29.ln1.weight
F32
F32
[1280]
v.blk.29.ln2.bias
F32
F32
[1280]
v.blk.29.ln2.weight
F32
F32
[1280]
v.blk.30
v.blk.30.attn_k.bias
F32
F32
[1280]
v.blk.30.attn_k.weight
F16
F16
[1280, 1280]
v.blk.30.attn_out.bias
F32
F32
[1280]
v.blk.30.attn_out.weight
F16
F16
[1280, 1280]
v.blk.30.attn_q.bias
F32
F32
[1280]
v.blk.30.attn_q.weight
F16
F16
[1280, 1280]
v.blk.30.attn_v.bias
F32
F32
[1280]
v.blk.30.attn_v.weight
F16
F16
[1280, 1280]
v.blk.30.ffn_down.bias
F32
F32
[1280]
v.blk.30.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.30.ffn_up.bias
F32
F32
[5120]
v.blk.30.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.30.ln1.bias
F32
F32
[1280]
v.blk.30.ln1.weight
F32
F32
[1280]
v.blk.30.ln2.bias
F32
F32
[1280]
v.blk.30.ln2.weight
F32
F32
[1280]
v.blk.31
v.blk.31.attn_k.bias
F32
F32
[1280]
v.blk.31.attn_k.weight
F16
F16
[1280, 1280]
v.blk.31.attn_out.bias
F32
F32
[1280]
v.blk.31.attn_out.weight
F16
F16
[1280, 1280]
v.blk.31.attn_q.bias
F32
F32
[1280]
v.blk.31.attn_q.weight
F16
F16
[1280, 1280]
v.blk.31.attn_v.bias
F32
F32
[1280]
v.blk.31.attn_v.weight
F16
F16
[1280, 1280]
v.blk.31.ffn_down.bias
F32
F32
[1280]
v.blk.31.ffn_down.weight
F16
F16
[5120, 1280]
v.blk.31.ffn_up.bias
F32
F32
[5120]
v.blk.31.ffn_up.weight
F16
F16
[1280, 5120]
v.blk.31.ln1.bias
F32
F32
[1280]
v.blk.31.ln1.weight
F32
F32
[1280]
v.blk.31.ln2.bias
F32
F32
[1280]
v.blk.31.ln2.weight
F32
F32
[1280]
v.cls_embd
F16
F16
[1280, 10]
v.patch_embd.weight
F16
F16
[768, 1280]
v.position_embd
F16
F16
[1280, 16384]
output_norm.weight
F32
F32
[2688]