nemotron3:33b-q8

111 20 hours ago

vision tools thinking audio 33b
c34ddc773b02 · 36GB
    Metadata
  • general.architecture
    nemotron_h_omni
  • general.file_type
    Q8_0
  • nemotron_h_omni.attention.head_count
    32
  • nemotron_h_omni.attention.head_count_kv
    [0, 0, 0, 0, 0, ...]
  • nemotron_h_omni.attention.key_length
    128
  • nemotron_h_omni.attention.layer_norm_epsilon
    1e-05
  • nemotron_h_omni.attention.layer_norm_rms_epsilon
    1e-05
  • nemotron_h_omni.attention.value_length
    128
  • nemotron_h_omni.audio.attention.head_count
    8
  • nemotron_h_omni.audio.attention.layer_norm_epsilon
    1e-05
  • nemotron_h_omni.audio.block_count
    24
  • nemotron_h_omni.audio.conv_kernel_size
    9
  • nemotron_h_omni.audio.embedding_length
    1024
  • nemotron_h_omni.audio.feed_forward_length
    4096
  • nemotron_h_omni.audio.num_mel_bins
    128
  • nemotron_h_omni.audio.projection_hidden_size
    4096
  • nemotron_h_omni.audio.sample_rate
    16000
  • nemotron_h_omni.audio.scale_input
    false
  • nemotron_h_omni.audio.sound_token_id
    27
  • nemotron_h_omni.audio.subsampling_conv_channels
    256
  • nemotron_h_omni.audio.subsampling_conv_kernel_size
    3
  • nemotron_h_omni.audio.subsampling_conv_stride
    2
  • nemotron_h_omni.audio.subsampling_factor
    8
  • nemotron_h_omni.block_count
    52
  • nemotron_h_omni.context_length
    131072
  • nemotron_h_omni.embedding_length
    2688
  • nemotron_h_omni.expert_count
    128
  • nemotron_h_omni.expert_feed_forward_length
    1856
  • nemotron_h_omni.expert_group_count
    1
  • nemotron_h_omni.expert_group_used_count
    1
  • nemotron_h_omni.expert_shared_count
    1
  • nemotron_h_omni.expert_shared_feed_forward_length
    3712
  • nemotron_h_omni.expert_used_count
    6
  • nemotron_h_omni.expert_weights_norm
    true
  • nemotron_h_omni.expert_weights_scale
    2.5
  • nemotron_h_omni.feed_forward_length
    [0, 1856, 0, 1856, 0, ...]
  • nemotron_h_omni.rope.dimension_count
    128
  • nemotron_h_omni.rope.freq_base
    10000
  • nemotron_h_omni.ssm.conv_kernel
    4
  • nemotron_h_omni.ssm.group_count
    8
  • nemotron_h_omni.ssm.inner_size
    4096
  • nemotron_h_omni.ssm.state_size
    128
  • nemotron_h_omni.ssm.time_step_rank
    64
  • nemotron_h_omni.vision.attention.head_count
    16
  • nemotron_h_omni.vision.attention.layer_norm_epsilon
    1e-06
  • nemotron_h_omni.vision.block_count
    32
  • nemotron_h_omni.vision.embedding_length
    1280
  • nemotron_h_omni.vision.feed_forward_length
    5120
  • nemotron_h_omni.vision.image_end_token_id
    20
  • nemotron_h_omni.vision.image_mean
    [0.48145467, 0.4578275, 0.40821072]
  • nemotron_h_omni.vision.image_size
    512
  • nemotron_h_omni.vision.image_start_token_id
    19
  • nemotron_h_omni.vision.image_std
    [0.26862955, 0.2613026, 0.2757771]
  • nemotron_h_omni.vision.image_token_id
    18
  • nemotron_h_omni.vision.max_num_patches
    13312
  • nemotron_h_omni.vision.max_tiles
    12
  • nemotron_h_omni.vision.min_num_patches
    1024
  • nemotron_h_omni.vision.num_channels
    3
  • nemotron_h_omni.vision.patch_size
    16
  • nemotron_h_omni.vision.projector.scale_factor
    2
  • nemotron_h_omni.vision.use_thumbnail
    true
  • tokenizer.ggml.add_bos_token
    false
  • tokenizer.ggml.add_eos_token
    false
  • tokenizer.ggml.add_unknown_token
    false
  • tokenizer.ggml.bos_token_id
    1
  • tokenizer.ggml.eos_token_id
    11
  • tokenizer.ggml.eos_token_ids
    [2, 11]
  • tokenizer.ggml.merges
    [Ġ Ġ, Ġ t, e r, i n, Ġ ĠĠĠ, ...]
  • tokenizer.ggml.model
    gpt2
  • tokenizer.ggml.pre
    default
  • tokenizer.ggml.scores
    [0, 1, 2, 3, 4, ...]
  • tokenizer.ggml.token_type
    [3, 3, 3, 3, 3, ...]
  • tokenizer.ggml.tokens
    [<unk>, <s>, </s>, [INST], [/INST], ...]
  • tokenizer.ggml.unknown_token_id
    0
  • Tensor
  • token_embd.weight
    Q8_0
    [2688, 131072]
  • a.blk.0
  • a.blk.0.attn_bias_u
    BF16
    [128, 8]
  • a.blk.0.attn_bias_v
    BF16
    [128, 8]
  • a.blk.0.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.0.attn_norm.bias
    F32
    [1024]
  • a.blk.0.attn_norm.weight
    F32
    [1024]
  • a.blk.0.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.0.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.0.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.0.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.0.conv_bn.bias
    F32
    [1024]
  • a.blk.0.conv_bn.running_mean
    F32
    [1024]
  • a.blk.0.conv_bn.running_var
    F32
    [1024]
  • a.blk.0.conv_bn.weight
    F32
    [1024]
  • a.blk.0.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.0.conv_norm.bias
    F32
    [1024]
  • a.blk.0.conv_norm.weight
    F32
    [1024]
  • a.blk.0.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.0.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.0.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.0.ffn1_norm.bias
    F32
    [1024]
  • a.blk.0.ffn1_norm.weight
    F32
    [1024]
  • a.blk.0.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.0.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.0.ffn2_norm.bias
    F32
    [1024]
  • a.blk.0.ffn2_norm.weight
    F32
    [1024]
  • a.blk.0.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.0.out_norm.bias
    F32
    [1024]
  • a.blk.0.out_norm.weight
    F32
    [1024]
  • a.blk.1
  • a.blk.1.attn_bias_u
    BF16
    [128, 8]
  • a.blk.1.attn_bias_v
    BF16
    [128, 8]
  • a.blk.1.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.1.attn_norm.bias
    F32
    [1024]
  • a.blk.1.attn_norm.weight
    F32
    [1024]
  • a.blk.1.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.1.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.1.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.1.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.1.conv_bn.bias
    F32
    [1024]
  • a.blk.1.conv_bn.running_mean
    F32
    [1024]
  • a.blk.1.conv_bn.running_var
    F32
    [1024]
  • a.blk.1.conv_bn.weight
    F32
    [1024]
  • a.blk.1.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.1.conv_norm.bias
    F32
    [1024]
  • a.blk.1.conv_norm.weight
    F32
    [1024]
  • a.blk.1.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.1.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.1.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.1.ffn1_norm.bias
    F32
    [1024]
  • a.blk.1.ffn1_norm.weight
    F32
    [1024]
  • a.blk.1.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.1.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.1.ffn2_norm.bias
    F32
    [1024]
  • a.blk.1.ffn2_norm.weight
    F32
    [1024]
  • a.blk.1.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.1.out_norm.bias
    F32
    [1024]
  • a.blk.1.out_norm.weight
    F32
    [1024]
  • a.blk.2
  • a.blk.2.attn_bias_u
    BF16
    [128, 8]
  • a.blk.2.attn_bias_v
    BF16
    [128, 8]
  • a.blk.2.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.2.attn_norm.bias
    F32
    [1024]
  • a.blk.2.attn_norm.weight
    F32
    [1024]
  • a.blk.2.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.2.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.2.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.2.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.2.conv_bn.bias
    F32
    [1024]
  • a.blk.2.conv_bn.running_mean
    F32
    [1024]
  • a.blk.2.conv_bn.running_var
    F32
    [1024]
  • a.blk.2.conv_bn.weight
    F32
    [1024]
  • a.blk.2.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.2.conv_norm.bias
    F32
    [1024]
  • a.blk.2.conv_norm.weight
    F32
    [1024]
  • a.blk.2.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.2.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.2.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.2.ffn1_norm.bias
    F32
    [1024]
  • a.blk.2.ffn1_norm.weight
    F32
    [1024]
  • a.blk.2.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.2.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.2.ffn2_norm.bias
    F32
    [1024]
  • a.blk.2.ffn2_norm.weight
    F32
    [1024]
  • a.blk.2.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.2.out_norm.bias
    F32
    [1024]
  • a.blk.2.out_norm.weight
    F32
    [1024]
  • a.blk.3
  • a.blk.3.attn_bias_u
    BF16
    [128, 8]
  • a.blk.3.attn_bias_v
    BF16
    [128, 8]
  • a.blk.3.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.3.attn_norm.bias
    F32
    [1024]
  • a.blk.3.attn_norm.weight
    F32
    [1024]
  • a.blk.3.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.3.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.3.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.3.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.3.conv_bn.bias
    F32
    [1024]
  • a.blk.3.conv_bn.running_mean
    F32
    [1024]
  • a.blk.3.conv_bn.running_var
    F32
    [1024]
  • a.blk.3.conv_bn.weight
    F32
    [1024]
  • a.blk.3.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.3.conv_norm.bias
    F32
    [1024]
  • a.blk.3.conv_norm.weight
    F32
    [1024]
  • a.blk.3.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.3.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.3.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.3.ffn1_norm.bias
    F32
    [1024]
  • a.blk.3.ffn1_norm.weight
    F32
    [1024]
  • a.blk.3.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.3.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.3.ffn2_norm.bias
    F32
    [1024]
  • a.blk.3.ffn2_norm.weight
    F32
    [1024]
  • a.blk.3.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.3.out_norm.bias
    F32
    [1024]
  • a.blk.3.out_norm.weight
    F32
    [1024]
  • a.blk.4
  • a.blk.4.attn_bias_u
    BF16
    [128, 8]
  • a.blk.4.attn_bias_v
    BF16
    [128, 8]
  • a.blk.4.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.4.attn_norm.bias
    F32
    [1024]
  • a.blk.4.attn_norm.weight
    F32
    [1024]
  • a.blk.4.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.4.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.4.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.4.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.4.conv_bn.bias
    F32
    [1024]
  • a.blk.4.conv_bn.running_mean
    F32
    [1024]
  • a.blk.4.conv_bn.running_var
    F32
    [1024]
  • a.blk.4.conv_bn.weight
    F32
    [1024]
  • a.blk.4.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.4.conv_norm.bias
    F32
    [1024]
  • a.blk.4.conv_norm.weight
    F32
    [1024]
  • a.blk.4.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.4.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.4.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.4.ffn1_norm.bias
    F32
    [1024]
  • a.blk.4.ffn1_norm.weight
    F32
    [1024]
  • a.blk.4.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.4.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.4.ffn2_norm.bias
    F32
    [1024]
  • a.blk.4.ffn2_norm.weight
    F32
    [1024]
  • a.blk.4.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.4.out_norm.bias
    F32
    [1024]
  • a.blk.4.out_norm.weight
    F32
    [1024]
  • a.blk.5
  • a.blk.5.attn_bias_u
    BF16
    [128, 8]
  • a.blk.5.attn_bias_v
    BF16
    [128, 8]
  • a.blk.5.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.5.attn_norm.bias
    F32
    [1024]
  • a.blk.5.attn_norm.weight
    F32
    [1024]
  • a.blk.5.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.5.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.5.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.5.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.5.conv_bn.bias
    F32
    [1024]
  • a.blk.5.conv_bn.running_mean
    F32
    [1024]
  • a.blk.5.conv_bn.running_var
    F32
    [1024]
  • a.blk.5.conv_bn.weight
    F32
    [1024]
  • a.blk.5.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.5.conv_norm.bias
    F32
    [1024]
  • a.blk.5.conv_norm.weight
    F32
    [1024]
  • a.blk.5.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.5.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.5.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.5.ffn1_norm.bias
    F32
    [1024]
  • a.blk.5.ffn1_norm.weight
    F32
    [1024]
  • a.blk.5.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.5.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.5.ffn2_norm.bias
    F32
    [1024]
  • a.blk.5.ffn2_norm.weight
    F32
    [1024]
  • a.blk.5.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.5.out_norm.bias
    F32
    [1024]
  • a.blk.5.out_norm.weight
    F32
    [1024]
  • a.blk.6
  • a.blk.6.attn_bias_u
    BF16
    [128, 8]
  • a.blk.6.attn_bias_v
    BF16
    [128, 8]
  • a.blk.6.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.6.attn_norm.bias
    F32
    [1024]
  • a.blk.6.attn_norm.weight
    F32
    [1024]
  • a.blk.6.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.6.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.6.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.6.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.6.conv_bn.bias
    F32
    [1024]
  • a.blk.6.conv_bn.running_mean
    F32
    [1024]
  • a.blk.6.conv_bn.running_var
    F32
    [1024]
  • a.blk.6.conv_bn.weight
    F32
    [1024]
  • a.blk.6.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.6.conv_norm.bias
    F32
    [1024]
  • a.blk.6.conv_norm.weight
    F32
    [1024]
  • a.blk.6.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.6.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.6.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.6.ffn1_norm.bias
    F32
    [1024]
  • a.blk.6.ffn1_norm.weight
    F32
    [1024]
  • a.blk.6.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.6.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.6.ffn2_norm.bias
    F32
    [1024]
  • a.blk.6.ffn2_norm.weight
    F32
    [1024]
  • a.blk.6.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.6.out_norm.bias
    F32
    [1024]
  • a.blk.6.out_norm.weight
    F32
    [1024]
  • a.blk.7
  • a.blk.7.attn_bias_u
    BF16
    [128, 8]
  • a.blk.7.attn_bias_v
    BF16
    [128, 8]
  • a.blk.7.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.7.attn_norm.bias
    F32
    [1024]
  • a.blk.7.attn_norm.weight
    F32
    [1024]
  • a.blk.7.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.7.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.7.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.7.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.7.conv_bn.bias
    F32
    [1024]
  • a.blk.7.conv_bn.running_mean
    F32
    [1024]
  • a.blk.7.conv_bn.running_var
    F32
    [1024]
  • a.blk.7.conv_bn.weight
    F32
    [1024]
  • a.blk.7.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.7.conv_norm.bias
    F32
    [1024]
  • a.blk.7.conv_norm.weight
    F32
    [1024]
  • a.blk.7.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.7.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.7.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.7.ffn1_norm.bias
    F32
    [1024]
  • a.blk.7.ffn1_norm.weight
    F32
    [1024]
  • a.blk.7.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.7.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.7.ffn2_norm.bias
    F32
    [1024]
  • a.blk.7.ffn2_norm.weight
    F32
    [1024]
  • a.blk.7.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.7.out_norm.bias
    F32
    [1024]
  • a.blk.7.out_norm.weight
    F32
    [1024]
  • a.blk.8
  • a.blk.8.attn_bias_u
    BF16
    [128, 8]
  • a.blk.8.attn_bias_v
    BF16
    [128, 8]
  • a.blk.8.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.8.attn_norm.bias
    F32
    [1024]
  • a.blk.8.attn_norm.weight
    F32
    [1024]
  • a.blk.8.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.8.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.8.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.8.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.8.conv_bn.bias
    F32
    [1024]
  • a.blk.8.conv_bn.running_mean
    F32
    [1024]
  • a.blk.8.conv_bn.running_var
    F32
    [1024]
  • a.blk.8.conv_bn.weight
    F32
    [1024]
  • a.blk.8.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.8.conv_norm.bias
    F32
    [1024]
  • a.blk.8.conv_norm.weight
    F32
    [1024]
  • a.blk.8.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.8.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.8.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.8.ffn1_norm.bias
    F32
    [1024]
  • a.blk.8.ffn1_norm.weight
    F32
    [1024]
  • a.blk.8.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.8.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.8.ffn2_norm.bias
    F32
    [1024]
  • a.blk.8.ffn2_norm.weight
    F32
    [1024]
  • a.blk.8.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.8.out_norm.bias
    F32
    [1024]
  • a.blk.8.out_norm.weight
    F32
    [1024]
  • a.blk.9
  • a.blk.9.attn_bias_u
    BF16
    [128, 8]
  • a.blk.9.attn_bias_v
    BF16
    [128, 8]
  • a.blk.9.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.9.attn_norm.bias
    F32
    [1024]
  • a.blk.9.attn_norm.weight
    F32
    [1024]
  • a.blk.9.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.9.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.9.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.9.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.9.conv_bn.bias
    F32
    [1024]
  • a.blk.9.conv_bn.running_mean
    F32
    [1024]
  • a.blk.9.conv_bn.running_var
    F32
    [1024]
  • a.blk.9.conv_bn.weight
    F32
    [1024]
  • a.blk.9.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.9.conv_norm.bias
    F32
    [1024]
  • a.blk.9.conv_norm.weight
    F32
    [1024]
  • a.blk.9.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.9.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.9.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.9.ffn1_norm.bias
    F32
    [1024]
  • a.blk.9.ffn1_norm.weight
    F32
    [1024]
  • a.blk.9.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.9.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.9.ffn2_norm.bias
    F32
    [1024]
  • a.blk.9.ffn2_norm.weight
    F32
    [1024]
  • a.blk.9.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.9.out_norm.bias
    F32
    [1024]
  • a.blk.9.out_norm.weight
    F32
    [1024]
  • a.blk.10
  • a.blk.10.attn_bias_u
    BF16
    [128, 8]
  • a.blk.10.attn_bias_v
    BF16
    [128, 8]
  • a.blk.10.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.10.attn_norm.bias
    F32
    [1024]
  • a.blk.10.attn_norm.weight
    F32
    [1024]
  • a.blk.10.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.10.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.10.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.10.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.10.conv_bn.bias
    F32
    [1024]
  • a.blk.10.conv_bn.running_mean
    F32
    [1024]
  • a.blk.10.conv_bn.running_var
    F32
    [1024]
  • a.blk.10.conv_bn.weight
    F32
    [1024]
  • a.blk.10.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.10.conv_norm.bias
    F32
    [1024]
  • a.blk.10.conv_norm.weight
    F32
    [1024]
  • a.blk.10.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.10.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.10.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.10.ffn1_norm.bias
    F32
    [1024]
  • a.blk.10.ffn1_norm.weight
    F32
    [1024]
  • a.blk.10.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.10.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.10.ffn2_norm.bias
    F32
    [1024]
  • a.blk.10.ffn2_norm.weight
    F32
    [1024]
  • a.blk.10.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.10.out_norm.bias
    F32
    [1024]
  • a.blk.10.out_norm.weight
    F32
    [1024]
  • a.blk.11
  • a.blk.11.attn_bias_u
    BF16
    [128, 8]
  • a.blk.11.attn_bias_v
    BF16
    [128, 8]
  • a.blk.11.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.11.attn_norm.bias
    F32
    [1024]
  • a.blk.11.attn_norm.weight
    F32
    [1024]
  • a.blk.11.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.11.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.11.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.11.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.11.conv_bn.bias
    F32
    [1024]
  • a.blk.11.conv_bn.running_mean
    F32
    [1024]
  • a.blk.11.conv_bn.running_var
    F32
    [1024]
  • a.blk.11.conv_bn.weight
    F32
    [1024]
  • a.blk.11.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.11.conv_norm.bias
    F32
    [1024]
  • a.blk.11.conv_norm.weight
    F32
    [1024]
  • a.blk.11.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.11.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.11.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.11.ffn1_norm.bias
    F32
    [1024]
  • a.blk.11.ffn1_norm.weight
    F32
    [1024]
  • a.blk.11.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.11.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.11.ffn2_norm.bias
    F32
    [1024]
  • a.blk.11.ffn2_norm.weight
    F32
    [1024]
  • a.blk.11.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.11.out_norm.bias
    F32
    [1024]
  • a.blk.11.out_norm.weight
    F32
    [1024]
  • a.blk.12
  • a.blk.12.attn_bias_u
    BF16
    [128, 8]
  • a.blk.12.attn_bias_v
    BF16
    [128, 8]
  • a.blk.12.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.12.attn_norm.bias
    F32
    [1024]
  • a.blk.12.attn_norm.weight
    F32
    [1024]
  • a.blk.12.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.12.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.12.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.12.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.12.conv_bn.bias
    F32
    [1024]
  • a.blk.12.conv_bn.running_mean
    F32
    [1024]
  • a.blk.12.conv_bn.running_var
    F32
    [1024]
  • a.blk.12.conv_bn.weight
    F32
    [1024]
  • a.blk.12.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.12.conv_norm.bias
    F32
    [1024]
  • a.blk.12.conv_norm.weight
    F32
    [1024]
  • a.blk.12.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.12.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.12.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.12.ffn1_norm.bias
    F32
    [1024]
  • a.blk.12.ffn1_norm.weight
    F32
    [1024]
  • a.blk.12.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.12.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.12.ffn2_norm.bias
    F32
    [1024]
  • a.blk.12.ffn2_norm.weight
    F32
    [1024]
  • a.blk.12.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.12.out_norm.bias
    F32
    [1024]
  • a.blk.12.out_norm.weight
    F32
    [1024]
  • a.blk.13
  • a.blk.13.attn_bias_u
    BF16
    [128, 8]
  • a.blk.13.attn_bias_v
    BF16
    [128, 8]
  • a.blk.13.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.13.attn_norm.bias
    F32
    [1024]
  • a.blk.13.attn_norm.weight
    F32
    [1024]
  • a.blk.13.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.13.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.13.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.13.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.13.conv_bn.bias
    F32
    [1024]
  • a.blk.13.conv_bn.running_mean
    F32
    [1024]
  • a.blk.13.conv_bn.running_var
    F32
    [1024]
  • a.blk.13.conv_bn.weight
    F32
    [1024]
  • a.blk.13.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.13.conv_norm.bias
    F32
    [1024]
  • a.blk.13.conv_norm.weight
    F32
    [1024]
  • a.blk.13.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.13.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.13.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.13.ffn1_norm.bias
    F32
    [1024]
  • a.blk.13.ffn1_norm.weight
    F32
    [1024]
  • a.blk.13.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.13.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.13.ffn2_norm.bias
    F32
    [1024]
  • a.blk.13.ffn2_norm.weight
    F32
    [1024]
  • a.blk.13.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.13.out_norm.bias
    F32
    [1024]
  • a.blk.13.out_norm.weight
    F32
    [1024]
  • a.blk.14
  • a.blk.14.attn_bias_u
    BF16
    [128, 8]
  • a.blk.14.attn_bias_v
    BF16
    [128, 8]
  • a.blk.14.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.14.attn_norm.bias
    F32
    [1024]
  • a.blk.14.attn_norm.weight
    F32
    [1024]
  • a.blk.14.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.14.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.14.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.14.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.14.conv_bn.bias
    F32
    [1024]
  • a.blk.14.conv_bn.running_mean
    F32
    [1024]
  • a.blk.14.conv_bn.running_var
    F32
    [1024]
  • a.blk.14.conv_bn.weight
    F32
    [1024]
  • a.blk.14.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.14.conv_norm.bias
    F32
    [1024]
  • a.blk.14.conv_norm.weight
    F32
    [1024]
  • a.blk.14.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.14.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.14.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.14.ffn1_norm.bias
    F32
    [1024]
  • a.blk.14.ffn1_norm.weight
    F32
    [1024]
  • a.blk.14.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.14.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.14.ffn2_norm.bias
    F32
    [1024]
  • a.blk.14.ffn2_norm.weight
    F32
    [1024]
  • a.blk.14.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.14.out_norm.bias
    F32
    [1024]
  • a.blk.14.out_norm.weight
    F32
    [1024]
  • a.blk.15
  • a.blk.15.attn_bias_u
    BF16
    [128, 8]
  • a.blk.15.attn_bias_v
    BF16
    [128, 8]
  • a.blk.15.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.15.attn_norm.bias
    F32
    [1024]
  • a.blk.15.attn_norm.weight
    F32
    [1024]
  • a.blk.15.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.15.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.15.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.15.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.15.conv_bn.bias
    F32
    [1024]
  • a.blk.15.conv_bn.running_mean
    F32
    [1024]
  • a.blk.15.conv_bn.running_var
    F32
    [1024]
  • a.blk.15.conv_bn.weight
    F32
    [1024]
  • a.blk.15.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.15.conv_norm.bias
    F32
    [1024]
  • a.blk.15.conv_norm.weight
    F32
    [1024]
  • a.blk.15.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.15.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.15.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.15.ffn1_norm.bias
    F32
    [1024]
  • a.blk.15.ffn1_norm.weight
    F32
    [1024]
  • a.blk.15.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.15.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.15.ffn2_norm.bias
    F32
    [1024]
  • a.blk.15.ffn2_norm.weight
    F32
    [1024]
  • a.blk.15.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.15.out_norm.bias
    F32
    [1024]
  • a.blk.15.out_norm.weight
    F32
    [1024]
  • a.blk.16
  • a.blk.16.attn_bias_u
    BF16
    [128, 8]
  • a.blk.16.attn_bias_v
    BF16
    [128, 8]
  • a.blk.16.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.16.attn_norm.bias
    F32
    [1024]
  • a.blk.16.attn_norm.weight
    F32
    [1024]
  • a.blk.16.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.16.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.16.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.16.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.16.conv_bn.bias
    F32
    [1024]
  • a.blk.16.conv_bn.running_mean
    F32
    [1024]
  • a.blk.16.conv_bn.running_var
    F32
    [1024]
  • a.blk.16.conv_bn.weight
    F32
    [1024]
  • a.blk.16.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.16.conv_norm.bias
    F32
    [1024]
  • a.blk.16.conv_norm.weight
    F32
    [1024]
  • a.blk.16.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.16.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.16.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.16.ffn1_norm.bias
    F32
    [1024]
  • a.blk.16.ffn1_norm.weight
    F32
    [1024]
  • a.blk.16.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.16.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.16.ffn2_norm.bias
    F32
    [1024]
  • a.blk.16.ffn2_norm.weight
    F32
    [1024]
  • a.blk.16.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.16.out_norm.bias
    F32
    [1024]
  • a.blk.16.out_norm.weight
    F32
    [1024]
  • a.blk.17
  • a.blk.17.attn_bias_u
    BF16
    [128, 8]
  • a.blk.17.attn_bias_v
    BF16
    [128, 8]
  • a.blk.17.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.17.attn_norm.bias
    F32
    [1024]
  • a.blk.17.attn_norm.weight
    F32
    [1024]
  • a.blk.17.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.17.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.17.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.17.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.17.conv_bn.bias
    F32
    [1024]
  • a.blk.17.conv_bn.running_mean
    F32
    [1024]
  • a.blk.17.conv_bn.running_var
    F32
    [1024]
  • a.blk.17.conv_bn.weight
    F32
    [1024]
  • a.blk.17.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.17.conv_norm.bias
    F32
    [1024]
  • a.blk.17.conv_norm.weight
    F32
    [1024]
  • a.blk.17.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.17.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.17.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.17.ffn1_norm.bias
    F32
    [1024]
  • a.blk.17.ffn1_norm.weight
    F32
    [1024]
  • a.blk.17.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.17.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.17.ffn2_norm.bias
    F32
    [1024]
  • a.blk.17.ffn2_norm.weight
    F32
    [1024]
  • a.blk.17.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.17.out_norm.bias
    F32
    [1024]
  • a.blk.17.out_norm.weight
    F32
    [1024]
  • a.blk.18
  • a.blk.18.attn_bias_u
    BF16
    [128, 8]
  • a.blk.18.attn_bias_v
    BF16
    [128, 8]
  • a.blk.18.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.18.attn_norm.bias
    F32
    [1024]
  • a.blk.18.attn_norm.weight
    F32
    [1024]
  • a.blk.18.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.18.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.18.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.18.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.18.conv_bn.bias
    F32
    [1024]
  • a.blk.18.conv_bn.running_mean
    F32
    [1024]
  • a.blk.18.conv_bn.running_var
    F32
    [1024]
  • a.blk.18.conv_bn.weight
    F32
    [1024]
  • a.blk.18.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.18.conv_norm.bias
    F32
    [1024]
  • a.blk.18.conv_norm.weight
    F32
    [1024]
  • a.blk.18.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.18.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.18.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.18.ffn1_norm.bias
    F32
    [1024]
  • a.blk.18.ffn1_norm.weight
    F32
    [1024]
  • a.blk.18.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.18.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.18.ffn2_norm.bias
    F32
    [1024]
  • a.blk.18.ffn2_norm.weight
    F32
    [1024]
  • a.blk.18.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.18.out_norm.bias
    F32
    [1024]
  • a.blk.18.out_norm.weight
    F32
    [1024]
  • a.blk.19
  • a.blk.19.attn_bias_u
    BF16
    [128, 8]
  • a.blk.19.attn_bias_v
    BF16
    [128, 8]
  • a.blk.19.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.19.attn_norm.bias
    F32
    [1024]
  • a.blk.19.attn_norm.weight
    F32
    [1024]
  • a.blk.19.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.19.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.19.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.19.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.19.conv_bn.bias
    F32
    [1024]
  • a.blk.19.conv_bn.running_mean
    F32
    [1024]
  • a.blk.19.conv_bn.running_var
    F32
    [1024]
  • a.blk.19.conv_bn.weight
    F32
    [1024]
  • a.blk.19.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.19.conv_norm.bias
    F32
    [1024]
  • a.blk.19.conv_norm.weight
    F32
    [1024]
  • a.blk.19.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.19.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.19.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.19.ffn1_norm.bias
    F32
    [1024]
  • a.blk.19.ffn1_norm.weight
    F32
    [1024]
  • a.blk.19.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.19.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.19.ffn2_norm.bias
    F32
    [1024]
  • a.blk.19.ffn2_norm.weight
    F32
    [1024]
  • a.blk.19.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.19.out_norm.bias
    F32
    [1024]
  • a.blk.19.out_norm.weight
    F32
    [1024]
  • a.blk.20
  • a.blk.20.attn_bias_u
    BF16
    [128, 8]
  • a.blk.20.attn_bias_v
    BF16
    [128, 8]
  • a.blk.20.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.20.attn_norm.bias
    F32
    [1024]
  • a.blk.20.attn_norm.weight
    F32
    [1024]
  • a.blk.20.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.20.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.20.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.20.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.20.conv_bn.bias
    F32
    [1024]
  • a.blk.20.conv_bn.running_mean
    F32
    [1024]
  • a.blk.20.conv_bn.running_var
    F32
    [1024]
  • a.blk.20.conv_bn.weight
    F32
    [1024]
  • a.blk.20.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.20.conv_norm.bias
    F32
    [1024]
  • a.blk.20.conv_norm.weight
    F32
    [1024]
  • a.blk.20.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.20.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.20.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.20.ffn1_norm.bias
    F32
    [1024]
  • a.blk.20.ffn1_norm.weight
    F32
    [1024]
  • a.blk.20.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.20.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.20.ffn2_norm.bias
    F32
    [1024]
  • a.blk.20.ffn2_norm.weight
    F32
    [1024]
  • a.blk.20.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.20.out_norm.bias
    F32
    [1024]
  • a.blk.20.out_norm.weight
    F32
    [1024]
  • a.blk.21
  • a.blk.21.attn_bias_u
    BF16
    [128, 8]
  • a.blk.21.attn_bias_v
    BF16
    [128, 8]
  • a.blk.21.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.21.attn_norm.bias
    F32
    [1024]
  • a.blk.21.attn_norm.weight
    F32
    [1024]
  • a.blk.21.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.21.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.21.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.21.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.21.conv_bn.bias
    F32
    [1024]
  • a.blk.21.conv_bn.running_mean
    F32
    [1024]
  • a.blk.21.conv_bn.running_var
    F32
    [1024]
  • a.blk.21.conv_bn.weight
    F32
    [1024]
  • a.blk.21.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.21.conv_norm.bias
    F32
    [1024]
  • a.blk.21.conv_norm.weight
    F32
    [1024]
  • a.blk.21.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.21.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.21.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.21.ffn1_norm.bias
    F32
    [1024]
  • a.blk.21.ffn1_norm.weight
    F32
    [1024]
  • a.blk.21.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.21.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.21.ffn2_norm.bias
    F32
    [1024]
  • a.blk.21.ffn2_norm.weight
    F32
    [1024]
  • a.blk.21.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.21.out_norm.bias
    F32
    [1024]
  • a.blk.21.out_norm.weight
    F32
    [1024]
  • a.blk.22
  • a.blk.22.attn_bias_u
    BF16
    [128, 8]
  • a.blk.22.attn_bias_v
    BF16
    [128, 8]
  • a.blk.22.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.22.attn_norm.bias
    F32
    [1024]
  • a.blk.22.attn_norm.weight
    F32
    [1024]
  • a.blk.22.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.22.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.22.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.22.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.22.conv_bn.bias
    F32
    [1024]
  • a.blk.22.conv_bn.running_mean
    F32
    [1024]
  • a.blk.22.conv_bn.running_var
    F32
    [1024]
  • a.blk.22.conv_bn.weight
    F32
    [1024]
  • a.blk.22.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.22.conv_norm.bias
    F32
    [1024]
  • a.blk.22.conv_norm.weight
    F32
    [1024]
  • a.blk.22.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.22.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.22.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.22.ffn1_norm.bias
    F32
    [1024]
  • a.blk.22.ffn1_norm.weight
    F32
    [1024]
  • a.blk.22.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.22.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.22.ffn2_norm.bias
    F32
    [1024]
  • a.blk.22.ffn2_norm.weight
    F32
    [1024]
  • a.blk.22.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.22.out_norm.bias
    F32
    [1024]
  • a.blk.22.out_norm.weight
    F32
    [1024]
  • a.blk.23
  • a.blk.23.attn_bias_u
    BF16
    [128, 8]
  • a.blk.23.attn_bias_v
    BF16
    [128, 8]
  • a.blk.23.attn_k.weight
    BF16
    [1024, 1024]
  • a.blk.23.attn_norm.bias
    F32
    [1024]
  • a.blk.23.attn_norm.weight
    F32
    [1024]
  • a.blk.23.attn_out.weight
    BF16
    [1024, 1024]
  • a.blk.23.attn_q.weight
    BF16
    [1024, 1024]
  • a.blk.23.attn_rel_k.weight
    BF16
    [1024, 1024]
  • a.blk.23.attn_v.weight
    BF16
    [1024, 1024]
  • a.blk.23.conv_bn.bias
    F32
    [1024]
  • a.blk.23.conv_bn.running_mean
    F32
    [1024]
  • a.blk.23.conv_bn.running_var
    F32
    [1024]
  • a.blk.23.conv_bn.weight
    F32
    [1024]
  • a.blk.23.conv_dw.weight
    F32
    [9, 1024]
  • a.blk.23.conv_norm.bias
    F32
    [1024]
  • a.blk.23.conv_norm.weight
    F32
    [1024]
  • a.blk.23.conv_pw1.weight
    BF16
    [1024, 2048]
  • a.blk.23.conv_pw2.weight
    BF16
    [1024, 1024]
  • a.blk.23.ffn1_down.weight
    BF16
    [4096, 1024]
  • a.blk.23.ffn1_norm.bias
    F32
    [1024]
  • a.blk.23.ffn1_norm.weight
    F32
    [1024]
  • a.blk.23.ffn1_up.weight
    BF16
    [1024, 4096]
  • a.blk.23.ffn2_down.weight
    BF16
    [4096, 1024]
  • a.blk.23.ffn2_norm.bias
    F32
    [1024]
  • a.blk.23.ffn2_norm.weight
    F32
    [1024]
  • a.blk.23.ffn2_up.weight
    BF16
    [1024, 4096]
  • a.blk.23.out_norm.bias
    F32
    [1024]
  • a.blk.23.out_norm.weight
    F32
    [1024]
  • a.feature_extractor.fb
    F32
    [257, 128, 1]
  • a.feature_extractor.window
    F32
    [400]
  • a.subsampling.conv0.bias
    F32
    [256]
  • a.subsampling.conv0.weight
    F32
    [3, 3, 1, 256]
  • a.subsampling.dw1.bias
    F32
    [256]
  • a.subsampling.dw1.weight
    F32
    [3, 3, 1, 256]
  • a.subsampling.dw2.bias
    F32
    [256]
  • a.subsampling.dw2.weight
    F32
    [3, 3, 1, 256]
  • a.subsampling.linear.bias
    F32
    [1024]
  • a.subsampling.linear.weight
    F32
    [4096, 1024]
  • a.subsampling.pw1.bias
    F32
    [256]
  • a.subsampling.pw1.weight
    F32
    [1, 1, 256, 256]
  • a.subsampling.pw2.bias
    F32
    [256]
  • a.subsampling.pw2.weight
    F32
    [1, 1, 256, 256]
  • blk.0
  • blk.0.attn_norm.weight
    F32
    [2688]
  • blk.0.ssm_a
    F32
    [1, 64]
  • blk.0.ssm_conv1d.bias
    F32
    [6144]
  • blk.0.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.0.ssm_d
    F32
    [1, 64]
  • blk.0.ssm_dt.bias
    F32
    [64]
  • blk.0.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.0.ssm_norm.weight
    F32
    [512, 8]
  • blk.0.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.1
  • blk.1.attn_norm.weight
    F32
    [2688]
  • blk.1.exp_probs_b.bias
    F32
    [128]
  • blk.1.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.1.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.1.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.1.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.1.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.2
  • blk.2.attn_norm.weight
    F32
    [2688]
  • blk.2.ssm_a
    F32
    [1, 64]
  • blk.2.ssm_conv1d.bias
    F32
    [6144]
  • blk.2.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.2.ssm_d
    F32
    [1, 64]
  • blk.2.ssm_dt.bias
    F32
    [64]
  • blk.2.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.2.ssm_norm.weight
    F32
    [512, 8]
  • blk.2.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.3
  • blk.3.attn_norm.weight
    F32
    [2688]
  • blk.3.exp_probs_b.bias
    F32
    [128]
  • blk.3.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.3.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.3.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.3.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.3.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.4
  • blk.4.attn_norm.weight
    F32
    [2688]
  • blk.4.ssm_a
    F32
    [1, 64]
  • blk.4.ssm_conv1d.bias
    F32
    [6144]
  • blk.4.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.4.ssm_d
    F32
    [1, 64]
  • blk.4.ssm_dt.bias
    F32
    [64]
  • blk.4.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.4.ssm_norm.weight
    F32
    [512, 8]
  • blk.4.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.5
  • blk.5.attn_k.weight
    Q8_0
    [2688, 256]
  • blk.5.attn_norm.weight
    F32
    [2688]
  • blk.5.attn_output.weight
    Q8_0
    [4096, 2688]
  • blk.5.attn_q.weight
    Q8_0
    [2688, 4096]
  • blk.5.attn_v.weight
    Q8_0
    [2688, 256]
  • blk.6
  • blk.6.attn_norm.weight
    F32
    [2688]
  • blk.6.exp_probs_b.bias
    F32
    [128]
  • blk.6.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.6.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.6.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.6.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.6.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.7
  • blk.7.attn_norm.weight
    F32
    [2688]
  • blk.7.ssm_a
    F32
    [1, 64]
  • blk.7.ssm_conv1d.bias
    F32
    [6144]
  • blk.7.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.7.ssm_d
    F32
    [1, 64]
  • blk.7.ssm_dt.bias
    F32
    [64]
  • blk.7.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.7.ssm_norm.weight
    F32
    [512, 8]
  • blk.7.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.8
  • blk.8.attn_norm.weight
    F32
    [2688]
  • blk.8.exp_probs_b.bias
    F32
    [128]
  • blk.8.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.8.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.8.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.8.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.8.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.9
  • blk.9.attn_norm.weight
    F32
    [2688]
  • blk.9.ssm_a
    F32
    [1, 64]
  • blk.9.ssm_conv1d.bias
    F32
    [6144]
  • blk.9.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.9.ssm_d
    F32
    [1, 64]
  • blk.9.ssm_dt.bias
    F32
    [64]
  • blk.9.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.9.ssm_norm.weight
    F32
    [512, 8]
  • blk.9.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.10
  • blk.10.attn_norm.weight
    F32
    [2688]
  • blk.10.exp_probs_b.bias
    F32
    [128]
  • blk.10.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.10.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.10.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.10.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.10.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.11
  • blk.11.attn_norm.weight
    F32
    [2688]
  • blk.11.ssm_a
    F32
    [1, 64]
  • blk.11.ssm_conv1d.bias
    F32
    [6144]
  • blk.11.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.11.ssm_d
    F32
    [1, 64]
  • blk.11.ssm_dt.bias
    F32
    [64]
  • blk.11.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.11.ssm_norm.weight
    F32
    [512, 8]
  • blk.11.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.12
  • blk.12.attn_k.weight
    Q8_0
    [2688, 256]
  • blk.12.attn_norm.weight
    F32
    [2688]
  • blk.12.attn_output.weight
    Q8_0
    [4096, 2688]
  • blk.12.attn_q.weight
    Q8_0
    [2688, 4096]
  • blk.12.attn_v.weight
    Q8_0
    [2688, 256]
  • blk.13
  • blk.13.attn_norm.weight
    F32
    [2688]
  • blk.13.exp_probs_b.bias
    F32
    [128]
  • blk.13.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.13.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.13.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.13.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.13.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.14
  • blk.14.attn_norm.weight
    F32
    [2688]
  • blk.14.ssm_a
    F32
    [1, 64]
  • blk.14.ssm_conv1d.bias
    F32
    [6144]
  • blk.14.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.14.ssm_d
    F32
    [1, 64]
  • blk.14.ssm_dt.bias
    F32
    [64]
  • blk.14.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.14.ssm_norm.weight
    F32
    [512, 8]
  • blk.14.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.15
  • blk.15.attn_norm.weight
    F32
    [2688]
  • blk.15.exp_probs_b.bias
    F32
    [128]
  • blk.15.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.15.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.15.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.15.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.15.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.16
  • blk.16.attn_norm.weight
    F32
    [2688]
  • blk.16.ssm_a
    F32
    [1, 64]
  • blk.16.ssm_conv1d.bias
    F32
    [6144]
  • blk.16.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.16.ssm_d
    F32
    [1, 64]
  • blk.16.ssm_dt.bias
    F32
    [64]
  • blk.16.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.16.ssm_norm.weight
    F32
    [512, 8]
  • blk.16.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.17
  • blk.17.attn_norm.weight
    F32
    [2688]
  • blk.17.exp_probs_b.bias
    F32
    [128]
  • blk.17.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.17.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.17.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.17.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.17.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.18
  • blk.18.attn_norm.weight
    F32
    [2688]
  • blk.18.ssm_a
    F32
    [1, 64]
  • blk.18.ssm_conv1d.bias
    F32
    [6144]
  • blk.18.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.18.ssm_d
    F32
    [1, 64]
  • blk.18.ssm_dt.bias
    F32
    [64]
  • blk.18.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.18.ssm_norm.weight
    F32
    [512, 8]
  • blk.18.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.19
  • blk.19.attn_k.weight
    Q8_0
    [2688, 256]
  • blk.19.attn_norm.weight
    F32
    [2688]
  • blk.19.attn_output.weight
    Q8_0
    [4096, 2688]
  • blk.19.attn_q.weight
    Q8_0
    [2688, 4096]
  • blk.19.attn_v.weight
    Q8_0
    [2688, 256]
  • blk.20
  • blk.20.attn_norm.weight
    F32
    [2688]
  • blk.20.exp_probs_b.bias
    F32
    [128]
  • blk.20.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.20.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.20.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.20.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.20.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.21
  • blk.21.attn_norm.weight
    F32
    [2688]
  • blk.21.ssm_a
    F32
    [1, 64]
  • blk.21.ssm_conv1d.bias
    F32
    [6144]
  • blk.21.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.21.ssm_d
    F32
    [1, 64]
  • blk.21.ssm_dt.bias
    F32
    [64]
  • blk.21.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.21.ssm_norm.weight
    F32
    [512, 8]
  • blk.21.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.22
  • blk.22.attn_norm.weight
    F32
    [2688]
  • blk.22.exp_probs_b.bias
    F32
    [128]
  • blk.22.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.22.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.22.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.22.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.22.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.23
  • blk.23.attn_norm.weight
    F32
    [2688]
  • blk.23.ssm_a
    F32
    [1, 64]
  • blk.23.ssm_conv1d.bias
    F32
    [6144]
  • blk.23.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.23.ssm_d
    F32
    [1, 64]
  • blk.23.ssm_dt.bias
    F32
    [64]
  • blk.23.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.23.ssm_norm.weight
    F32
    [512, 8]
  • blk.23.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.24
  • blk.24.attn_norm.weight
    F32
    [2688]
  • blk.24.exp_probs_b.bias
    F32
    [128]
  • blk.24.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.24.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.24.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.24.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.24.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.25
  • blk.25.attn_norm.weight
    F32
    [2688]
  • blk.25.ssm_a
    F32
    [1, 64]
  • blk.25.ssm_conv1d.bias
    F32
    [6144]
  • blk.25.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.25.ssm_d
    F32
    [1, 64]
  • blk.25.ssm_dt.bias
    F32
    [64]
  • blk.25.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.25.ssm_norm.weight
    F32
    [512, 8]
  • blk.25.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.26
  • blk.26.attn_k.weight
    Q8_0
    [2688, 256]
  • blk.26.attn_norm.weight
    F32
    [2688]
  • blk.26.attn_output.weight
    Q8_0
    [4096, 2688]
  • blk.26.attn_q.weight
    Q8_0
    [2688, 4096]
  • blk.26.attn_v.weight
    Q8_0
    [2688, 256]
  • blk.27
  • blk.27.attn_norm.weight
    F32
    [2688]
  • blk.27.exp_probs_b.bias
    F32
    [128]
  • blk.27.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.27.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.27.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.27.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.27.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.28
  • blk.28.attn_norm.weight
    F32
    [2688]
  • blk.28.ssm_a
    F32
    [1, 64]
  • blk.28.ssm_conv1d.bias
    F32
    [6144]
  • blk.28.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.28.ssm_d
    F32
    [1, 64]
  • blk.28.ssm_dt.bias
    F32
    [64]
  • blk.28.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.28.ssm_norm.weight
    F32
    [512, 8]
  • blk.28.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.29
  • blk.29.attn_norm.weight
    F32
    [2688]
  • blk.29.exp_probs_b.bias
    F32
    [128]
  • blk.29.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.29.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.29.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.29.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.29.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.30
  • blk.30.attn_norm.weight
    F32
    [2688]
  • blk.30.ssm_a
    F32
    [1, 64]
  • blk.30.ssm_conv1d.bias
    F32
    [6144]
  • blk.30.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.30.ssm_d
    F32
    [1, 64]
  • blk.30.ssm_dt.bias
    F32
    [64]
  • blk.30.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.30.ssm_norm.weight
    F32
    [512, 8]
  • blk.30.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.31
  • blk.31.attn_norm.weight
    F32
    [2688]
  • blk.31.exp_probs_b.bias
    F32
    [128]
  • blk.31.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.31.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.31.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.31.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.31.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.32
  • blk.32.attn_norm.weight
    F32
    [2688]
  • blk.32.ssm_a
    F32
    [1, 64]
  • blk.32.ssm_conv1d.bias
    F32
    [6144]
  • blk.32.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.32.ssm_d
    F32
    [1, 64]
  • blk.32.ssm_dt.bias
    F32
    [64]
  • blk.32.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.32.ssm_norm.weight
    F32
    [512, 8]
  • blk.32.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.33
  • blk.33.attn_k.weight
    Q8_0
    [2688, 256]
  • blk.33.attn_norm.weight
    F32
    [2688]
  • blk.33.attn_output.weight
    Q8_0
    [4096, 2688]
  • blk.33.attn_q.weight
    Q8_0
    [2688, 4096]
  • blk.33.attn_v.weight
    Q8_0
    [2688, 256]
  • blk.34
  • blk.34.attn_norm.weight
    F32
    [2688]
  • blk.34.exp_probs_b.bias
    F32
    [128]
  • blk.34.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.34.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.34.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.34.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.34.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.35
  • blk.35.attn_norm.weight
    F32
    [2688]
  • blk.35.ssm_a
    F32
    [1, 64]
  • blk.35.ssm_conv1d.bias
    F32
    [6144]
  • blk.35.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.35.ssm_d
    F32
    [1, 64]
  • blk.35.ssm_dt.bias
    F32
    [64]
  • blk.35.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.35.ssm_norm.weight
    F32
    [512, 8]
  • blk.35.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.36
  • blk.36.attn_norm.weight
    F32
    [2688]
  • blk.36.exp_probs_b.bias
    F32
    [128]
  • blk.36.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.36.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.36.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.36.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.36.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.37
  • blk.37.attn_norm.weight
    F32
    [2688]
  • blk.37.ssm_a
    F32
    [1, 64]
  • blk.37.ssm_conv1d.bias
    F32
    [6144]
  • blk.37.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.37.ssm_d
    F32
    [1, 64]
  • blk.37.ssm_dt.bias
    F32
    [64]
  • blk.37.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.37.ssm_norm.weight
    F32
    [512, 8]
  • blk.37.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.38
  • blk.38.attn_norm.weight
    F32
    [2688]
  • blk.38.exp_probs_b.bias
    F32
    [128]
  • blk.38.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.38.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.38.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.38.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.38.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.39
  • blk.39.attn_norm.weight
    F32
    [2688]
  • blk.39.ssm_a
    F32
    [1, 64]
  • blk.39.ssm_conv1d.bias
    F32
    [6144]
  • blk.39.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.39.ssm_d
    F32
    [1, 64]
  • blk.39.ssm_dt.bias
    F32
    [64]
  • blk.39.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.39.ssm_norm.weight
    F32
    [512, 8]
  • blk.39.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.40
  • blk.40.attn_norm.weight
    F32
    [2688]
  • blk.40.exp_probs_b.bias
    F32
    [128]
  • blk.40.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.40.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.40.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.40.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.40.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.41
  • blk.41.attn_norm.weight
    F32
    [2688]
  • blk.41.ssm_a
    F32
    [1, 64]
  • blk.41.ssm_conv1d.bias
    F32
    [6144]
  • blk.41.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.41.ssm_d
    F32
    [1, 64]
  • blk.41.ssm_dt.bias
    F32
    [64]
  • blk.41.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.41.ssm_norm.weight
    F32
    [512, 8]
  • blk.41.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.42
  • blk.42.attn_k.weight
    Q8_0
    [2688, 256]
  • blk.42.attn_norm.weight
    F32
    [2688]
  • blk.42.attn_output.weight
    Q8_0
    [4096, 2688]
  • blk.42.attn_q.weight
    Q8_0
    [2688, 4096]
  • blk.42.attn_v.weight
    Q8_0
    [2688, 256]
  • blk.43
  • blk.43.attn_norm.weight
    F32
    [2688]
  • blk.43.exp_probs_b.bias
    F32
    [128]
  • blk.43.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.43.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.43.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.43.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.43.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.44
  • blk.44.attn_norm.weight
    F32
    [2688]
  • blk.44.ssm_a
    F32
    [1, 64]
  • blk.44.ssm_conv1d.bias
    F32
    [6144]
  • blk.44.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.44.ssm_d
    F32
    [1, 64]
  • blk.44.ssm_dt.bias
    F32
    [64]
  • blk.44.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.44.ssm_norm.weight
    F32
    [512, 8]
  • blk.44.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.45
  • blk.45.attn_norm.weight
    F32
    [2688]
  • blk.45.exp_probs_b.bias
    F32
    [128]
  • blk.45.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.45.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.45.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.45.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.45.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.46
  • blk.46.attn_norm.weight
    F32
    [2688]
  • blk.46.ssm_a
    F32
    [1, 64]
  • blk.46.ssm_conv1d.bias
    F32
    [6144]
  • blk.46.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.46.ssm_d
    F32
    [1, 64]
  • blk.46.ssm_dt.bias
    F32
    [64]
  • blk.46.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.46.ssm_norm.weight
    F32
    [512, 8]
  • blk.46.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.47
  • blk.47.attn_norm.weight
    F32
    [2688]
  • blk.47.exp_probs_b.bias
    F32
    [128]
  • blk.47.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.47.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.47.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.47.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.47.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.48
  • blk.48.attn_norm.weight
    F32
    [2688]
  • blk.48.ssm_a
    F32
    [1, 64]
  • blk.48.ssm_conv1d.bias
    F32
    [6144]
  • blk.48.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.48.ssm_d
    F32
    [1, 64]
  • blk.48.ssm_dt.bias
    F32
    [64]
  • blk.48.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.48.ssm_norm.weight
    F32
    [512, 8]
  • blk.48.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.49
  • blk.49.attn_norm.weight
    F32
    [2688]
  • blk.49.exp_probs_b.bias
    F32
    [128]
  • blk.49.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.49.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.49.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.49.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.49.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • blk.50
  • blk.50.attn_norm.weight
    F32
    [2688]
  • blk.50.ssm_a
    F32
    [1, 64]
  • blk.50.ssm_conv1d.bias
    F32
    [6144]
  • blk.50.ssm_conv1d.weight
    F32
    [4, 6144]
  • blk.50.ssm_d
    F32
    [1, 64]
  • blk.50.ssm_dt.bias
    F32
    [64]
  • blk.50.ssm_in.weight
    Q8_0
    [2688, 10304]
  • blk.50.ssm_norm.weight
    F32
    [512, 8]
  • blk.50.ssm_out.weight
    Q8_0
    [4096, 2688]
  • blk.51
  • blk.51.attn_norm.weight
    F32
    [2688]
  • blk.51.exp_probs_b.bias
    F32
    [128]
  • blk.51.ffn_down_exps.weight
    Q8_0
    [1856, 2688, 128]
  • blk.51.ffn_down_shexp.weight
    Q8_0
    [3712, 2688]
  • blk.51.ffn_gate_inp.weight
    F32
    [2688, 128]
  • blk.51.ffn_up_exps.weight
    Q8_0
    [2688, 1856, 128]
  • blk.51.ffn_up_shexp.weight
    Q8_0
    [2688, 3712]
  • mm.1.weight
    F16
    [5120, 20480]
  • mm.2.weight
    F16
    [20480, 2688]
  • mm.a.1.weight
    F16
    [1024, 4096]
  • mm.a.2.weight
    F16
    [4096, 2688]
  • mm.a.norm.weight
    F32
    [1024]
  • mm.norm.weight
    F32
    [5120]
  • output.weight
    Q8_0
    [2688, 131072]
  • v.blk.0
  • v.blk.0.attn_k.bias
    F32
    [1280]
  • v.blk.0.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.0.attn_out.bias
    F32
    [1280]
  • v.blk.0.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.0.attn_q.bias
    F32
    [1280]
  • v.blk.0.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.0.attn_v.bias
    F32
    [1280]
  • v.blk.0.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.0.ffn_down.bias
    F32
    [1280]
  • v.blk.0.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.0.ffn_up.bias
    F32
    [5120]
  • v.blk.0.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.0.ln1.bias
    F32
    [1280]
  • v.blk.0.ln1.weight
    F32
    [1280]
  • v.blk.0.ln2.bias
    F32
    [1280]
  • v.blk.0.ln2.weight
    F32
    [1280]
  • v.blk.1
  • v.blk.1.attn_k.bias
    F32
    [1280]
  • v.blk.1.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.1.attn_out.bias
    F32
    [1280]
  • v.blk.1.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.1.attn_q.bias
    F32
    [1280]
  • v.blk.1.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.1.attn_v.bias
    F32
    [1280]
  • v.blk.1.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.1.ffn_down.bias
    F32
    [1280]
  • v.blk.1.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.1.ffn_up.bias
    F32
    [5120]
  • v.blk.1.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.1.ln1.bias
    F32
    [1280]
  • v.blk.1.ln1.weight
    F32
    [1280]
  • v.blk.1.ln2.bias
    F32
    [1280]
  • v.blk.1.ln2.weight
    F32
    [1280]
  • v.blk.2
  • v.blk.2.attn_k.bias
    F32
    [1280]
  • v.blk.2.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.2.attn_out.bias
    F32
    [1280]
  • v.blk.2.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.2.attn_q.bias
    F32
    [1280]
  • v.blk.2.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.2.attn_v.bias
    F32
    [1280]
  • v.blk.2.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.2.ffn_down.bias
    F32
    [1280]
  • v.blk.2.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.2.ffn_up.bias
    F32
    [5120]
  • v.blk.2.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.2.ln1.bias
    F32
    [1280]
  • v.blk.2.ln1.weight
    F32
    [1280]
  • v.blk.2.ln2.bias
    F32
    [1280]
  • v.blk.2.ln2.weight
    F32
    [1280]
  • v.blk.3
  • v.blk.3.attn_k.bias
    F32
    [1280]
  • v.blk.3.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.3.attn_out.bias
    F32
    [1280]
  • v.blk.3.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.3.attn_q.bias
    F32
    [1280]
  • v.blk.3.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.3.attn_v.bias
    F32
    [1280]
  • v.blk.3.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.3.ffn_down.bias
    F32
    [1280]
  • v.blk.3.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.3.ffn_up.bias
    F32
    [5120]
  • v.blk.3.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.3.ln1.bias
    F32
    [1280]
  • v.blk.3.ln1.weight
    F32
    [1280]
  • v.blk.3.ln2.bias
    F32
    [1280]
  • v.blk.3.ln2.weight
    F32
    [1280]
  • v.blk.4
  • v.blk.4.attn_k.bias
    F32
    [1280]
  • v.blk.4.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.4.attn_out.bias
    F32
    [1280]
  • v.blk.4.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.4.attn_q.bias
    F32
    [1280]
  • v.blk.4.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.4.attn_v.bias
    F32
    [1280]
  • v.blk.4.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.4.ffn_down.bias
    F32
    [1280]
  • v.blk.4.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.4.ffn_up.bias
    F32
    [5120]
  • v.blk.4.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.4.ln1.bias
    F32
    [1280]
  • v.blk.4.ln1.weight
    F32
    [1280]
  • v.blk.4.ln2.bias
    F32
    [1280]
  • v.blk.4.ln2.weight
    F32
    [1280]
  • v.blk.5
  • v.blk.5.attn_k.bias
    F32
    [1280]
  • v.blk.5.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.5.attn_out.bias
    F32
    [1280]
  • v.blk.5.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.5.attn_q.bias
    F32
    [1280]
  • v.blk.5.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.5.attn_v.bias
    F32
    [1280]
  • v.blk.5.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.5.ffn_down.bias
    F32
    [1280]
  • v.blk.5.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.5.ffn_up.bias
    F32
    [5120]
  • v.blk.5.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.5.ln1.bias
    F32
    [1280]
  • v.blk.5.ln1.weight
    F32
    [1280]
  • v.blk.5.ln2.bias
    F32
    [1280]
  • v.blk.5.ln2.weight
    F32
    [1280]
  • v.blk.6
  • v.blk.6.attn_k.bias
    F32
    [1280]
  • v.blk.6.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.6.attn_out.bias
    F32
    [1280]
  • v.blk.6.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.6.attn_q.bias
    F32
    [1280]
  • v.blk.6.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.6.attn_v.bias
    F32
    [1280]
  • v.blk.6.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.6.ffn_down.bias
    F32
    [1280]
  • v.blk.6.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.6.ffn_up.bias
    F32
    [5120]
  • v.blk.6.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.6.ln1.bias
    F32
    [1280]
  • v.blk.6.ln1.weight
    F32
    [1280]
  • v.blk.6.ln2.bias
    F32
    [1280]
  • v.blk.6.ln2.weight
    F32
    [1280]
  • v.blk.7
  • v.blk.7.attn_k.bias
    F32
    [1280]
  • v.blk.7.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.7.attn_out.bias
    F32
    [1280]
  • v.blk.7.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.7.attn_q.bias
    F32
    [1280]
  • v.blk.7.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.7.attn_v.bias
    F32
    [1280]
  • v.blk.7.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.7.ffn_down.bias
    F32
    [1280]
  • v.blk.7.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.7.ffn_up.bias
    F32
    [5120]
  • v.blk.7.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.7.ln1.bias
    F32
    [1280]
  • v.blk.7.ln1.weight
    F32
    [1280]
  • v.blk.7.ln2.bias
    F32
    [1280]
  • v.blk.7.ln2.weight
    F32
    [1280]
  • v.blk.8
  • v.blk.8.attn_k.bias
    F32
    [1280]
  • v.blk.8.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.8.attn_out.bias
    F32
    [1280]
  • v.blk.8.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.8.attn_q.bias
    F32
    [1280]
  • v.blk.8.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.8.attn_v.bias
    F32
    [1280]
  • v.blk.8.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.8.ffn_down.bias
    F32
    [1280]
  • v.blk.8.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.8.ffn_up.bias
    F32
    [5120]
  • v.blk.8.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.8.ln1.bias
    F32
    [1280]
  • v.blk.8.ln1.weight
    F32
    [1280]
  • v.blk.8.ln2.bias
    F32
    [1280]
  • v.blk.8.ln2.weight
    F32
    [1280]
  • v.blk.9
  • v.blk.9.attn_k.bias
    F32
    [1280]
  • v.blk.9.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.9.attn_out.bias
    F32
    [1280]
  • v.blk.9.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.9.attn_q.bias
    F32
    [1280]
  • v.blk.9.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.9.attn_v.bias
    F32
    [1280]
  • v.blk.9.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.9.ffn_down.bias
    F32
    [1280]
  • v.blk.9.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.9.ffn_up.bias
    F32
    [5120]
  • v.blk.9.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.9.ln1.bias
    F32
    [1280]
  • v.blk.9.ln1.weight
    F32
    [1280]
  • v.blk.9.ln2.bias
    F32
    [1280]
  • v.blk.9.ln2.weight
    F32
    [1280]
  • v.blk.10
  • v.blk.10.attn_k.bias
    F32
    [1280]
  • v.blk.10.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.10.attn_out.bias
    F32
    [1280]
  • v.blk.10.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.10.attn_q.bias
    F32
    [1280]
  • v.blk.10.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.10.attn_v.bias
    F32
    [1280]
  • v.blk.10.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.10.ffn_down.bias
    F32
    [1280]
  • v.blk.10.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.10.ffn_up.bias
    F32
    [5120]
  • v.blk.10.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.10.ln1.bias
    F32
    [1280]
  • v.blk.10.ln1.weight
    F32
    [1280]
  • v.blk.10.ln2.bias
    F32
    [1280]
  • v.blk.10.ln2.weight
    F32
    [1280]
  • v.blk.11
  • v.blk.11.attn_k.bias
    F32
    [1280]
  • v.blk.11.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.11.attn_out.bias
    F32
    [1280]
  • v.blk.11.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.11.attn_q.bias
    F32
    [1280]
  • v.blk.11.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.11.attn_v.bias
    F32
    [1280]
  • v.blk.11.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.11.ffn_down.bias
    F32
    [1280]
  • v.blk.11.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.11.ffn_up.bias
    F32
    [5120]
  • v.blk.11.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.11.ln1.bias
    F32
    [1280]
  • v.blk.11.ln1.weight
    F32
    [1280]
  • v.blk.11.ln2.bias
    F32
    [1280]
  • v.blk.11.ln2.weight
    F32
    [1280]
  • v.blk.12
  • v.blk.12.attn_k.bias
    F32
    [1280]
  • v.blk.12.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.12.attn_out.bias
    F32
    [1280]
  • v.blk.12.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.12.attn_q.bias
    F32
    [1280]
  • v.blk.12.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.12.attn_v.bias
    F32
    [1280]
  • v.blk.12.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.12.ffn_down.bias
    F32
    [1280]
  • v.blk.12.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.12.ffn_up.bias
    F32
    [5120]
  • v.blk.12.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.12.ln1.bias
    F32
    [1280]
  • v.blk.12.ln1.weight
    F32
    [1280]
  • v.blk.12.ln2.bias
    F32
    [1280]
  • v.blk.12.ln2.weight
    F32
    [1280]
  • v.blk.13
  • v.blk.13.attn_k.bias
    F32
    [1280]
  • v.blk.13.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.13.attn_out.bias
    F32
    [1280]
  • v.blk.13.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.13.attn_q.bias
    F32
    [1280]
  • v.blk.13.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.13.attn_v.bias
    F32
    [1280]
  • v.blk.13.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.13.ffn_down.bias
    F32
    [1280]
  • v.blk.13.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.13.ffn_up.bias
    F32
    [5120]
  • v.blk.13.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.13.ln1.bias
    F32
    [1280]
  • v.blk.13.ln1.weight
    F32
    [1280]
  • v.blk.13.ln2.bias
    F32
    [1280]
  • v.blk.13.ln2.weight
    F32
    [1280]
  • v.blk.14
  • v.blk.14.attn_k.bias
    F32
    [1280]
  • v.blk.14.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.14.attn_out.bias
    F32
    [1280]
  • v.blk.14.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.14.attn_q.bias
    F32
    [1280]
  • v.blk.14.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.14.attn_v.bias
    F32
    [1280]
  • v.blk.14.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.14.ffn_down.bias
    F32
    [1280]
  • v.blk.14.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.14.ffn_up.bias
    F32
    [5120]
  • v.blk.14.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.14.ln1.bias
    F32
    [1280]
  • v.blk.14.ln1.weight
    F32
    [1280]
  • v.blk.14.ln2.bias
    F32
    [1280]
  • v.blk.14.ln2.weight
    F32
    [1280]
  • v.blk.15
  • v.blk.15.attn_k.bias
    F32
    [1280]
  • v.blk.15.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.15.attn_out.bias
    F32
    [1280]
  • v.blk.15.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.15.attn_q.bias
    F32
    [1280]
  • v.blk.15.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.15.attn_v.bias
    F32
    [1280]
  • v.blk.15.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.15.ffn_down.bias
    F32
    [1280]
  • v.blk.15.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.15.ffn_up.bias
    F32
    [5120]
  • v.blk.15.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.15.ln1.bias
    F32
    [1280]
  • v.blk.15.ln1.weight
    F32
    [1280]
  • v.blk.15.ln2.bias
    F32
    [1280]
  • v.blk.15.ln2.weight
    F32
    [1280]
  • v.blk.16
  • v.blk.16.attn_k.bias
    F32
    [1280]
  • v.blk.16.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.16.attn_out.bias
    F32
    [1280]
  • v.blk.16.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.16.attn_q.bias
    F32
    [1280]
  • v.blk.16.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.16.attn_v.bias
    F32
    [1280]
  • v.blk.16.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.16.ffn_down.bias
    F32
    [1280]
  • v.blk.16.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.16.ffn_up.bias
    F32
    [5120]
  • v.blk.16.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.16.ln1.bias
    F32
    [1280]
  • v.blk.16.ln1.weight
    F32
    [1280]
  • v.blk.16.ln2.bias
    F32
    [1280]
  • v.blk.16.ln2.weight
    F32
    [1280]
  • v.blk.17
  • v.blk.17.attn_k.bias
    F32
    [1280]
  • v.blk.17.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.17.attn_out.bias
    F32
    [1280]
  • v.blk.17.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.17.attn_q.bias
    F32
    [1280]
  • v.blk.17.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.17.attn_v.bias
    F32
    [1280]
  • v.blk.17.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.17.ffn_down.bias
    F32
    [1280]
  • v.blk.17.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.17.ffn_up.bias
    F32
    [5120]
  • v.blk.17.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.17.ln1.bias
    F32
    [1280]
  • v.blk.17.ln1.weight
    F32
    [1280]
  • v.blk.17.ln2.bias
    F32
    [1280]
  • v.blk.17.ln2.weight
    F32
    [1280]
  • v.blk.18
  • v.blk.18.attn_k.bias
    F32
    [1280]
  • v.blk.18.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.18.attn_out.bias
    F32
    [1280]
  • v.blk.18.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.18.attn_q.bias
    F32
    [1280]
  • v.blk.18.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.18.attn_v.bias
    F32
    [1280]
  • v.blk.18.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.18.ffn_down.bias
    F32
    [1280]
  • v.blk.18.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.18.ffn_up.bias
    F32
    [5120]
  • v.blk.18.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.18.ln1.bias
    F32
    [1280]
  • v.blk.18.ln1.weight
    F32
    [1280]
  • v.blk.18.ln2.bias
    F32
    [1280]
  • v.blk.18.ln2.weight
    F32
    [1280]
  • v.blk.19
  • v.blk.19.attn_k.bias
    F32
    [1280]
  • v.blk.19.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.19.attn_out.bias
    F32
    [1280]
  • v.blk.19.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.19.attn_q.bias
    F32
    [1280]
  • v.blk.19.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.19.attn_v.bias
    F32
    [1280]
  • v.blk.19.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.19.ffn_down.bias
    F32
    [1280]
  • v.blk.19.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.19.ffn_up.bias
    F32
    [5120]
  • v.blk.19.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.19.ln1.bias
    F32
    [1280]
  • v.blk.19.ln1.weight
    F32
    [1280]
  • v.blk.19.ln2.bias
    F32
    [1280]
  • v.blk.19.ln2.weight
    F32
    [1280]
  • v.blk.20
  • v.blk.20.attn_k.bias
    F32
    [1280]
  • v.blk.20.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.20.attn_out.bias
    F32
    [1280]
  • v.blk.20.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.20.attn_q.bias
    F32
    [1280]
  • v.blk.20.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.20.attn_v.bias
    F32
    [1280]
  • v.blk.20.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.20.ffn_down.bias
    F32
    [1280]
  • v.blk.20.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.20.ffn_up.bias
    F32
    [5120]
  • v.blk.20.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.20.ln1.bias
    F32
    [1280]
  • v.blk.20.ln1.weight
    F32
    [1280]
  • v.blk.20.ln2.bias
    F32
    [1280]
  • v.blk.20.ln2.weight
    F32
    [1280]
  • v.blk.21
  • v.blk.21.attn_k.bias
    F32
    [1280]
  • v.blk.21.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.21.attn_out.bias
    F32
    [1280]
  • v.blk.21.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.21.attn_q.bias
    F32
    [1280]
  • v.blk.21.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.21.attn_v.bias
    F32
    [1280]
  • v.blk.21.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.21.ffn_down.bias
    F32
    [1280]
  • v.blk.21.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.21.ffn_up.bias
    F32
    [5120]
  • v.blk.21.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.21.ln1.bias
    F32
    [1280]
  • v.blk.21.ln1.weight
    F32
    [1280]
  • v.blk.21.ln2.bias
    F32
    [1280]
  • v.blk.21.ln2.weight
    F32
    [1280]
  • v.blk.22
  • v.blk.22.attn_k.bias
    F32
    [1280]
  • v.blk.22.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.22.attn_out.bias
    F32
    [1280]
  • v.blk.22.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.22.attn_q.bias
    F32
    [1280]
  • v.blk.22.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.22.attn_v.bias
    F32
    [1280]
  • v.blk.22.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.22.ffn_down.bias
    F32
    [1280]
  • v.blk.22.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.22.ffn_up.bias
    F32
    [5120]
  • v.blk.22.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.22.ln1.bias
    F32
    [1280]
  • v.blk.22.ln1.weight
    F32
    [1280]
  • v.blk.22.ln2.bias
    F32
    [1280]
  • v.blk.22.ln2.weight
    F32
    [1280]
  • v.blk.23
  • v.blk.23.attn_k.bias
    F32
    [1280]
  • v.blk.23.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.23.attn_out.bias
    F32
    [1280]
  • v.blk.23.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.23.attn_q.bias
    F32
    [1280]
  • v.blk.23.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.23.attn_v.bias
    F32
    [1280]
  • v.blk.23.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.23.ffn_down.bias
    F32
    [1280]
  • v.blk.23.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.23.ffn_up.bias
    F32
    [5120]
  • v.blk.23.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.23.ln1.bias
    F32
    [1280]
  • v.blk.23.ln1.weight
    F32
    [1280]
  • v.blk.23.ln2.bias
    F32
    [1280]
  • v.blk.23.ln2.weight
    F32
    [1280]
  • v.blk.24
  • v.blk.24.attn_k.bias
    F32
    [1280]
  • v.blk.24.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.24.attn_out.bias
    F32
    [1280]
  • v.blk.24.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.24.attn_q.bias
    F32
    [1280]
  • v.blk.24.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.24.attn_v.bias
    F32
    [1280]
  • v.blk.24.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.24.ffn_down.bias
    F32
    [1280]
  • v.blk.24.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.24.ffn_up.bias
    F32
    [5120]
  • v.blk.24.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.24.ln1.bias
    F32
    [1280]
  • v.blk.24.ln1.weight
    F32
    [1280]
  • v.blk.24.ln2.bias
    F32
    [1280]
  • v.blk.24.ln2.weight
    F32
    [1280]
  • v.blk.25
  • v.blk.25.attn_k.bias
    F32
    [1280]
  • v.blk.25.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.25.attn_out.bias
    F32
    [1280]
  • v.blk.25.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.25.attn_q.bias
    F32
    [1280]
  • v.blk.25.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.25.attn_v.bias
    F32
    [1280]
  • v.blk.25.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.25.ffn_down.bias
    F32
    [1280]
  • v.blk.25.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.25.ffn_up.bias
    F32
    [5120]
  • v.blk.25.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.25.ln1.bias
    F32
    [1280]
  • v.blk.25.ln1.weight
    F32
    [1280]
  • v.blk.25.ln2.bias
    F32
    [1280]
  • v.blk.25.ln2.weight
    F32
    [1280]
  • v.blk.26
  • v.blk.26.attn_k.bias
    F32
    [1280]
  • v.blk.26.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.26.attn_out.bias
    F32
    [1280]
  • v.blk.26.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.26.attn_q.bias
    F32
    [1280]
  • v.blk.26.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.26.attn_v.bias
    F32
    [1280]
  • v.blk.26.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.26.ffn_down.bias
    F32
    [1280]
  • v.blk.26.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.26.ffn_up.bias
    F32
    [5120]
  • v.blk.26.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.26.ln1.bias
    F32
    [1280]
  • v.blk.26.ln1.weight
    F32
    [1280]
  • v.blk.26.ln2.bias
    F32
    [1280]
  • v.blk.26.ln2.weight
    F32
    [1280]
  • v.blk.27
  • v.blk.27.attn_k.bias
    F32
    [1280]
  • v.blk.27.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.27.attn_out.bias
    F32
    [1280]
  • v.blk.27.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.27.attn_q.bias
    F32
    [1280]
  • v.blk.27.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.27.attn_v.bias
    F32
    [1280]
  • v.blk.27.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.27.ffn_down.bias
    F32
    [1280]
  • v.blk.27.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.27.ffn_up.bias
    F32
    [5120]
  • v.blk.27.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.27.ln1.bias
    F32
    [1280]
  • v.blk.27.ln1.weight
    F32
    [1280]
  • v.blk.27.ln2.bias
    F32
    [1280]
  • v.blk.27.ln2.weight
    F32
    [1280]
  • v.blk.28
  • v.blk.28.attn_k.bias
    F32
    [1280]
  • v.blk.28.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.28.attn_out.bias
    F32
    [1280]
  • v.blk.28.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.28.attn_q.bias
    F32
    [1280]
  • v.blk.28.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.28.attn_v.bias
    F32
    [1280]
  • v.blk.28.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.28.ffn_down.bias
    F32
    [1280]
  • v.blk.28.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.28.ffn_up.bias
    F32
    [5120]
  • v.blk.28.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.28.ln1.bias
    F32
    [1280]
  • v.blk.28.ln1.weight
    F32
    [1280]
  • v.blk.28.ln2.bias
    F32
    [1280]
  • v.blk.28.ln2.weight
    F32
    [1280]
  • v.blk.29
  • v.blk.29.attn_k.bias
    F32
    [1280]
  • v.blk.29.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.29.attn_out.bias
    F32
    [1280]
  • v.blk.29.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.29.attn_q.bias
    F32
    [1280]
  • v.blk.29.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.29.attn_v.bias
    F32
    [1280]
  • v.blk.29.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.29.ffn_down.bias
    F32
    [1280]
  • v.blk.29.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.29.ffn_up.bias
    F32
    [5120]
  • v.blk.29.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.29.ln1.bias
    F32
    [1280]
  • v.blk.29.ln1.weight
    F32
    [1280]
  • v.blk.29.ln2.bias
    F32
    [1280]
  • v.blk.29.ln2.weight
    F32
    [1280]
  • v.blk.30
  • v.blk.30.attn_k.bias
    F32
    [1280]
  • v.blk.30.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.30.attn_out.bias
    F32
    [1280]
  • v.blk.30.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.30.attn_q.bias
    F32
    [1280]
  • v.blk.30.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.30.attn_v.bias
    F32
    [1280]
  • v.blk.30.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.30.ffn_down.bias
    F32
    [1280]
  • v.blk.30.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.30.ffn_up.bias
    F32
    [5120]
  • v.blk.30.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.30.ln1.bias
    F32
    [1280]
  • v.blk.30.ln1.weight
    F32
    [1280]
  • v.blk.30.ln2.bias
    F32
    [1280]
  • v.blk.30.ln2.weight
    F32
    [1280]
  • v.blk.31
  • v.blk.31.attn_k.bias
    F32
    [1280]
  • v.blk.31.attn_k.weight
    F16
    [1280, 1280]
  • v.blk.31.attn_out.bias
    F32
    [1280]
  • v.blk.31.attn_out.weight
    F16
    [1280, 1280]
  • v.blk.31.attn_q.bias
    F32
    [1280]
  • v.blk.31.attn_q.weight
    F16
    [1280, 1280]
  • v.blk.31.attn_v.bias
    F32
    [1280]
  • v.blk.31.attn_v.weight
    F16
    [1280, 1280]
  • v.blk.31.ffn_down.bias
    F32
    [1280]
  • v.blk.31.ffn_down.weight
    F16
    [5120, 1280]
  • v.blk.31.ffn_up.bias
    F32
    [5120]
  • v.blk.31.ffn_up.weight
    F16
    [1280, 5120]
  • v.blk.31.ln1.bias
    F32
    [1280]
  • v.blk.31.ln1.weight
    F32
    [1280]
  • v.blk.31.ln2.bias
    F32
    [1280]
  • v.blk.31.ln2.weight
    F32
    [1280]
  • v.cls_embd
    F16
    [1280, 10]
  • v.patch_embd.weight
    F16
    [768, 1280]
  • v.position_embd
    F16
    [1280, 16384]
  • output_norm.weight
    F32
    [2688]