nemotron-3-super:120b-a12b-q8_0

43 30 minutes ago

NVIDIA Nemotron 3 Super is a 120B open MoE model activating just 12B parameters to deliver maximum compute efficiency and accuracy for complex multi-agent applications.

tools thinking cloud 120b
7e614521d8bc · 132GB
    Metadata
  • general.architecture
    nemotron_h_moe
  • general.file_type
    Q8_0
  • nemotron_h_moe.attention.head_count
    32
  • nemotron_h_moe.attention.head_count_kv
    [0, 0, 0, 0, 0, ...]
  • nemotron_h_moe.attention.key_length
    128
  • nemotron_h_moe.attention.layer_norm_epsilon
    1e-05
  • nemotron_h_moe.attention.layer_norm_rms_epsilon
    1e-05
  • nemotron_h_moe.attention.value_length
    128
  • nemotron_h_moe.block_count
    88
  • nemotron_h_moe.context_length
    262144
  • nemotron_h_moe.embedding_length
    4096
  • nemotron_h_moe.expert_count
    512
  • nemotron_h_moe.expert_feed_forward_length
    2688
  • nemotron_h_moe.expert_group_count
    1
  • nemotron_h_moe.expert_group_used_count
    1
  • nemotron_h_moe.expert_shared_count
    1
  • nemotron_h_moe.expert_shared_feed_forward_length
    5376
  • nemotron_h_moe.expert_used_count
    22
  • nemotron_h_moe.expert_weights_norm
    true
  • nemotron_h_moe.expert_weights_scale
    5
  • nemotron_h_moe.feed_forward_length
    [0, 2688, 0, 2688, 0, ...]
  • nemotron_h_moe.rope.dimension_count
    128
  • nemotron_h_moe.rope.freq_base
    10000
  • nemotron_h_moe.ssm.conv_kernel
    4
  • nemotron_h_moe.ssm.group_count
    8
  • nemotron_h_moe.ssm.inner_size
    8192
  • nemotron_h_moe.ssm.state_size
    128
  • nemotron_h_moe.ssm.time_step_rank
    128
  • tokenizer.ggml.add_bos_token
    false
  • tokenizer.ggml.add_eos_token
    false
  • tokenizer.ggml.add_padding_token
    false
  • tokenizer.ggml.add_unknown_token
    false
  • tokenizer.ggml.bos_token_id
    1
  • tokenizer.ggml.eos_token_id
    11
  • tokenizer.ggml.eos_token_ids
    [2, 11]
  • tokenizer.ggml.merges
    [Ġ Ġ, Ġ t, e r, i n, Ġ ĠĠĠ, ...]
  • tokenizer.ggml.model
    gpt2
  • tokenizer.ggml.padding_token_id
    11
  • tokenizer.ggml.pre
    default
  • tokenizer.ggml.scores
    [0, 1, 2, 3, 4, ...]
  • tokenizer.ggml.token_type
    [3, 3, 3, 3, 3, ...]
  • tokenizer.ggml.tokens
    [<unk>, <s>, </s>, [INST], [/INST], ...]
  • tokenizer.ggml.unknown_token_id
    0
  • Tensor
  • token_embd.weight
    Q8_0
    [4096, 131072]
  • blk.0
  • blk.0.attn_norm.weight
    F32
    [4096]
  • blk.0.ssm_a
    F32
    [1, 128]
  • blk.0.ssm_conv1d.bias
    F32
    [10240]
  • blk.0.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.0.ssm_d
    F32
    [1, 128]
  • blk.0.ssm_dt.bias
    F32
    [128]
  • blk.0.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.0.ssm_norm.weight
    F32
    [1024, 8]
  • blk.0.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.1
  • blk.1.attn_norm.weight
    F32
    [4096]
  • blk.1.exp_probs_b.bias
    F32
    [512]
  • blk.1.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.1.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.1.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.1.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.1.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.1.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.1.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.2
  • blk.2.attn_norm.weight
    F32
    [4096]
  • blk.2.ssm_a
    F32
    [1, 128]
  • blk.2.ssm_conv1d.bias
    F32
    [10240]
  • blk.2.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.2.ssm_d
    F32
    [1, 128]
  • blk.2.ssm_dt.bias
    F32
    [128]
  • blk.2.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.2.ssm_norm.weight
    F32
    [1024, 8]
  • blk.2.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.3
  • blk.3.attn_norm.weight
    F32
    [4096]
  • blk.3.exp_probs_b.bias
    F32
    [512]
  • blk.3.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.3.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.3.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.3.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.3.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.3.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.3.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.4
  • blk.4.attn_norm.weight
    F32
    [4096]
  • blk.4.ssm_a
    F32
    [1, 128]
  • blk.4.ssm_conv1d.bias
    F32
    [10240]
  • blk.4.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.4.ssm_d
    F32
    [1, 128]
  • blk.4.ssm_dt.bias
    F32
    [128]
  • blk.4.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.4.ssm_norm.weight
    F32
    [1024, 8]
  • blk.4.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.5
  • blk.5.attn_norm.weight
    F32
    [4096]
  • blk.5.exp_probs_b.bias
    F32
    [512]
  • blk.5.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.5.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.5.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.5.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.5.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.5.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.5.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.6
  • blk.6.attn_norm.weight
    F32
    [4096]
  • blk.6.ssm_a
    F32
    [1, 128]
  • blk.6.ssm_conv1d.bias
    F32
    [10240]
  • blk.6.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.6.ssm_d
    F32
    [1, 128]
  • blk.6.ssm_dt.bias
    F32
    [128]
  • blk.6.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.6.ssm_norm.weight
    F32
    [1024, 8]
  • blk.6.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.7
  • blk.7.attn_k.weight
    Q8_0
    [4096, 256]
  • blk.7.attn_norm.weight
    F32
    [4096]
  • blk.7.attn_output.weight
    Q8_0
    [4096, 4096]
  • blk.7.attn_q.weight
    Q8_0
    [4096, 4096]
  • blk.7.attn_v.weight
    Q8_0
    [4096, 256]
  • blk.8
  • blk.8.attn_norm.weight
    F32
    [4096]
  • blk.8.exp_probs_b.bias
    F32
    [512]
  • blk.8.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.8.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.8.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.8.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.8.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.8.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.8.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.9
  • blk.9.attn_norm.weight
    F32
    [4096]
  • blk.9.ssm_a
    F32
    [1, 128]
  • blk.9.ssm_conv1d.bias
    F32
    [10240]
  • blk.9.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.9.ssm_d
    F32
    [1, 128]
  • blk.9.ssm_dt.bias
    F32
    [128]
  • blk.9.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.9.ssm_norm.weight
    F32
    [1024, 8]
  • blk.9.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.10
  • blk.10.attn_norm.weight
    F32
    [4096]
  • blk.10.exp_probs_b.bias
    F32
    [512]
  • blk.10.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.10.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.10.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.10.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.10.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.10.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.10.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.11
  • blk.11.attn_norm.weight
    F32
    [4096]
  • blk.11.ssm_a
    F32
    [1, 128]
  • blk.11.ssm_conv1d.bias
    F32
    [10240]
  • blk.11.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.11.ssm_d
    F32
    [1, 128]
  • blk.11.ssm_dt.bias
    F32
    [128]
  • blk.11.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.11.ssm_norm.weight
    F32
    [1024, 8]
  • blk.11.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.12
  • blk.12.attn_norm.weight
    F32
    [4096]
  • blk.12.exp_probs_b.bias
    F32
    [512]
  • blk.12.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.12.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.12.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.12.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.12.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.12.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.12.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.13
  • blk.13.attn_norm.weight
    F32
    [4096]
  • blk.13.ssm_a
    F32
    [1, 128]
  • blk.13.ssm_conv1d.bias
    F32
    [10240]
  • blk.13.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.13.ssm_d
    F32
    [1, 128]
  • blk.13.ssm_dt.bias
    F32
    [128]
  • blk.13.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.13.ssm_norm.weight
    F32
    [1024, 8]
  • blk.13.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.14
  • blk.14.attn_norm.weight
    F32
    [4096]
  • blk.14.exp_probs_b.bias
    F32
    [512]
  • blk.14.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.14.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.14.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.14.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.14.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.14.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.14.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.15
  • blk.15.attn_norm.weight
    F32
    [4096]
  • blk.15.ssm_a
    F32
    [1, 128]
  • blk.15.ssm_conv1d.bias
    F32
    [10240]
  • blk.15.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.15.ssm_d
    F32
    [1, 128]
  • blk.15.ssm_dt.bias
    F32
    [128]
  • blk.15.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.15.ssm_norm.weight
    F32
    [1024, 8]
  • blk.15.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.16
  • blk.16.attn_k.weight
    Q8_0
    [4096, 256]
  • blk.16.attn_norm.weight
    F32
    [4096]
  • blk.16.attn_output.weight
    Q8_0
    [4096, 4096]
  • blk.16.attn_q.weight
    Q8_0
    [4096, 4096]
  • blk.16.attn_v.weight
    Q8_0
    [4096, 256]
  • blk.17
  • blk.17.attn_norm.weight
    F32
    [4096]
  • blk.17.exp_probs_b.bias
    F32
    [512]
  • blk.17.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.17.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.17.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.17.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.17.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.17.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.17.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.18
  • blk.18.attn_norm.weight
    F32
    [4096]
  • blk.18.ssm_a
    F32
    [1, 128]
  • blk.18.ssm_conv1d.bias
    F32
    [10240]
  • blk.18.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.18.ssm_d
    F32
    [1, 128]
  • blk.18.ssm_dt.bias
    F32
    [128]
  • blk.18.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.18.ssm_norm.weight
    F32
    [1024, 8]
  • blk.18.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.19
  • blk.19.attn_norm.weight
    F32
    [4096]
  • blk.19.exp_probs_b.bias
    F32
    [512]
  • blk.19.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.19.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.19.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.19.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.19.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.19.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.19.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.20
  • blk.20.attn_norm.weight
    F32
    [4096]
  • blk.20.ssm_a
    F32
    [1, 128]
  • blk.20.ssm_conv1d.bias
    F32
    [10240]
  • blk.20.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.20.ssm_d
    F32
    [1, 128]
  • blk.20.ssm_dt.bias
    F32
    [128]
  • blk.20.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.20.ssm_norm.weight
    F32
    [1024, 8]
  • blk.20.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.21
  • blk.21.attn_norm.weight
    F32
    [4096]
  • blk.21.exp_probs_b.bias
    F32
    [512]
  • blk.21.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.21.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.21.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.21.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.21.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.21.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.21.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.22
  • blk.22.attn_norm.weight
    F32
    [4096]
  • blk.22.ssm_a
    F32
    [1, 128]
  • blk.22.ssm_conv1d.bias
    F32
    [10240]
  • blk.22.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.22.ssm_d
    F32
    [1, 128]
  • blk.22.ssm_dt.bias
    F32
    [128]
  • blk.22.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.22.ssm_norm.weight
    F32
    [1024, 8]
  • blk.22.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.23
  • blk.23.attn_norm.weight
    F32
    [4096]
  • blk.23.exp_probs_b.bias
    F32
    [512]
  • blk.23.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.23.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.23.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.23.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.23.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.23.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.23.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.24
  • blk.24.attn_norm.weight
    F32
    [4096]
  • blk.24.ssm_a
    F32
    [1, 128]
  • blk.24.ssm_conv1d.bias
    F32
    [10240]
  • blk.24.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.24.ssm_d
    F32
    [1, 128]
  • blk.24.ssm_dt.bias
    F32
    [128]
  • blk.24.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.24.ssm_norm.weight
    F32
    [1024, 8]
  • blk.24.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.25
  • blk.25.attn_k.weight
    Q8_0
    [4096, 256]
  • blk.25.attn_norm.weight
    F32
    [4096]
  • blk.25.attn_output.weight
    Q8_0
    [4096, 4096]
  • blk.25.attn_q.weight
    Q8_0
    [4096, 4096]
  • blk.25.attn_v.weight
    Q8_0
    [4096, 256]
  • blk.26
  • blk.26.attn_norm.weight
    F32
    [4096]
  • blk.26.exp_probs_b.bias
    F32
    [512]
  • blk.26.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.26.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.26.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.26.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.26.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.26.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.26.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.27
  • blk.27.attn_norm.weight
    F32
    [4096]
  • blk.27.ssm_a
    F32
    [1, 128]
  • blk.27.ssm_conv1d.bias
    F32
    [10240]
  • blk.27.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.27.ssm_d
    F32
    [1, 128]
  • blk.27.ssm_dt.bias
    F32
    [128]
  • blk.27.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.27.ssm_norm.weight
    F32
    [1024, 8]
  • blk.27.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.28
  • blk.28.attn_norm.weight
    F32
    [4096]
  • blk.28.exp_probs_b.bias
    F32
    [512]
  • blk.28.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.28.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.28.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.28.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.28.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.28.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.28.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.29
  • blk.29.attn_norm.weight
    F32
    [4096]
  • blk.29.ssm_a
    F32
    [1, 128]
  • blk.29.ssm_conv1d.bias
    F32
    [10240]
  • blk.29.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.29.ssm_d
    F32
    [1, 128]
  • blk.29.ssm_dt.bias
    F32
    [128]
  • blk.29.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.29.ssm_norm.weight
    F32
    [1024, 8]
  • blk.29.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.30
  • blk.30.attn_norm.weight
    F32
    [4096]
  • blk.30.exp_probs_b.bias
    F32
    [512]
  • blk.30.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.30.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.30.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.30.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.30.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.30.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.30.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.31
  • blk.31.attn_norm.weight
    F32
    [4096]
  • blk.31.ssm_a
    F32
    [1, 128]
  • blk.31.ssm_conv1d.bias
    F32
    [10240]
  • blk.31.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.31.ssm_d
    F32
    [1, 128]
  • blk.31.ssm_dt.bias
    F32
    [128]
  • blk.31.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.31.ssm_norm.weight
    F32
    [1024, 8]
  • blk.31.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.32
  • blk.32.attn_norm.weight
    F32
    [4096]
  • blk.32.exp_probs_b.bias
    F32
    [512]
  • blk.32.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.32.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.32.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.32.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.32.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.32.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.32.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.33
  • blk.33.attn_norm.weight
    F32
    [4096]
  • blk.33.ssm_a
    F32
    [1, 128]
  • blk.33.ssm_conv1d.bias
    F32
    [10240]
  • blk.33.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.33.ssm_d
    F32
    [1, 128]
  • blk.33.ssm_dt.bias
    F32
    [128]
  • blk.33.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.33.ssm_norm.weight
    F32
    [1024, 8]
  • blk.33.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.34
  • blk.34.attn_norm.weight
    F32
    [4096]
  • blk.34.exp_probs_b.bias
    F32
    [512]
  • blk.34.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.34.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.34.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.34.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.34.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.34.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.34.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.35
  • blk.35.attn_norm.weight
    F32
    [4096]
  • blk.35.ssm_a
    F32
    [1, 128]
  • blk.35.ssm_conv1d.bias
    F32
    [10240]
  • blk.35.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.35.ssm_d
    F32
    [1, 128]
  • blk.35.ssm_dt.bias
    F32
    [128]
  • blk.35.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.35.ssm_norm.weight
    F32
    [1024, 8]
  • blk.35.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.36
  • blk.36.attn_k.weight
    Q8_0
    [4096, 256]
  • blk.36.attn_norm.weight
    F32
    [4096]
  • blk.36.attn_output.weight
    Q8_0
    [4096, 4096]
  • blk.36.attn_q.weight
    Q8_0
    [4096, 4096]
  • blk.36.attn_v.weight
    Q8_0
    [4096, 256]
  • blk.37
  • blk.37.attn_norm.weight
    F32
    [4096]
  • blk.37.exp_probs_b.bias
    F32
    [512]
  • blk.37.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.37.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.37.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.37.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.37.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.37.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.37.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.38
  • blk.38.attn_norm.weight
    F32
    [4096]
  • blk.38.ssm_a
    F32
    [1, 128]
  • blk.38.ssm_conv1d.bias
    F32
    [10240]
  • blk.38.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.38.ssm_d
    F32
    [1, 128]
  • blk.38.ssm_dt.bias
    F32
    [128]
  • blk.38.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.38.ssm_norm.weight
    F32
    [1024, 8]
  • blk.38.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.39
  • blk.39.attn_norm.weight
    F32
    [4096]
  • blk.39.exp_probs_b.bias
    F32
    [512]
  • blk.39.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.39.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.39.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.39.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.39.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.39.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.39.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.40
  • blk.40.attn_norm.weight
    F32
    [4096]
  • blk.40.ssm_a
    F32
    [1, 128]
  • blk.40.ssm_conv1d.bias
    F32
    [10240]
  • blk.40.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.40.ssm_d
    F32
    [1, 128]
  • blk.40.ssm_dt.bias
    F32
    [128]
  • blk.40.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.40.ssm_norm.weight
    F32
    [1024, 8]
  • blk.40.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.41
  • blk.41.attn_norm.weight
    F32
    [4096]
  • blk.41.exp_probs_b.bias
    F32
    [512]
  • blk.41.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.41.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.41.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.41.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.41.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.41.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.41.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.42
  • blk.42.attn_norm.weight
    F32
    [4096]
  • blk.42.ssm_a
    F32
    [1, 128]
  • blk.42.ssm_conv1d.bias
    F32
    [10240]
  • blk.42.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.42.ssm_d
    F32
    [1, 128]
  • blk.42.ssm_dt.bias
    F32
    [128]
  • blk.42.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.42.ssm_norm.weight
    F32
    [1024, 8]
  • blk.42.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.43
  • blk.43.attn_norm.weight
    F32
    [4096]
  • blk.43.exp_probs_b.bias
    F32
    [512]
  • blk.43.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.43.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.43.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.43.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.43.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.43.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.43.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.44
  • blk.44.attn_norm.weight
    F32
    [4096]
  • blk.44.ssm_a
    F32
    [1, 128]
  • blk.44.ssm_conv1d.bias
    F32
    [10240]
  • blk.44.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.44.ssm_d
    F32
    [1, 128]
  • blk.44.ssm_dt.bias
    F32
    [128]
  • blk.44.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.44.ssm_norm.weight
    F32
    [1024, 8]
  • blk.44.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.45
  • blk.45.attn_norm.weight
    F32
    [4096]
  • blk.45.exp_probs_b.bias
    F32
    [512]
  • blk.45.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.45.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.45.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.45.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.45.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.45.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.45.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.46
  • blk.46.attn_norm.weight
    F32
    [4096]
  • blk.46.ssm_a
    F32
    [1, 128]
  • blk.46.ssm_conv1d.bias
    F32
    [10240]
  • blk.46.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.46.ssm_d
    F32
    [1, 128]
  • blk.46.ssm_dt.bias
    F32
    [128]
  • blk.46.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.46.ssm_norm.weight
    F32
    [1024, 8]
  • blk.46.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.47
  • blk.47.attn_k.weight
    Q8_0
    [4096, 256]
  • blk.47.attn_norm.weight
    F32
    [4096]
  • blk.47.attn_output.weight
    Q8_0
    [4096, 4096]
  • blk.47.attn_q.weight
    Q8_0
    [4096, 4096]
  • blk.47.attn_v.weight
    Q8_0
    [4096, 256]
  • blk.48
  • blk.48.attn_norm.weight
    F32
    [4096]
  • blk.48.exp_probs_b.bias
    F32
    [512]
  • blk.48.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.48.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.48.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.48.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.48.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.48.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.48.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.49
  • blk.49.attn_norm.weight
    F32
    [4096]
  • blk.49.ssm_a
    F32
    [1, 128]
  • blk.49.ssm_conv1d.bias
    F32
    [10240]
  • blk.49.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.49.ssm_d
    F32
    [1, 128]
  • blk.49.ssm_dt.bias
    F32
    [128]
  • blk.49.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.49.ssm_norm.weight
    F32
    [1024, 8]
  • blk.49.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.50
  • blk.50.attn_norm.weight
    F32
    [4096]
  • blk.50.exp_probs_b.bias
    F32
    [512]
  • blk.50.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.50.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.50.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.50.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.50.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.50.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.50.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.51
  • blk.51.attn_norm.weight
    F32
    [4096]
  • blk.51.ssm_a
    F32
    [1, 128]
  • blk.51.ssm_conv1d.bias
    F32
    [10240]
  • blk.51.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.51.ssm_d
    F32
    [1, 128]
  • blk.51.ssm_dt.bias
    F32
    [128]
  • blk.51.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.51.ssm_norm.weight
    F32
    [1024, 8]
  • blk.51.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.52
  • blk.52.attn_norm.weight
    F32
    [4096]
  • blk.52.exp_probs_b.bias
    F32
    [512]
  • blk.52.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.52.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.52.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.52.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.52.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.52.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.52.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.53
  • blk.53.attn_norm.weight
    F32
    [4096]
  • blk.53.ssm_a
    F32
    [1, 128]
  • blk.53.ssm_conv1d.bias
    F32
    [10240]
  • blk.53.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.53.ssm_d
    F32
    [1, 128]
  • blk.53.ssm_dt.bias
    F32
    [128]
  • blk.53.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.53.ssm_norm.weight
    F32
    [1024, 8]
  • blk.53.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.54
  • blk.54.attn_norm.weight
    F32
    [4096]
  • blk.54.exp_probs_b.bias
    F32
    [512]
  • blk.54.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.54.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.54.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.54.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.54.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.54.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.54.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.55
  • blk.55.attn_norm.weight
    F32
    [4096]
  • blk.55.ssm_a
    F32
    [1, 128]
  • blk.55.ssm_conv1d.bias
    F32
    [10240]
  • blk.55.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.55.ssm_d
    F32
    [1, 128]
  • blk.55.ssm_dt.bias
    F32
    [128]
  • blk.55.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.55.ssm_norm.weight
    F32
    [1024, 8]
  • blk.55.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.56
  • blk.56.attn_norm.weight
    F32
    [4096]
  • blk.56.exp_probs_b.bias
    F32
    [512]
  • blk.56.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.56.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.56.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.56.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.56.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.56.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.56.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.57
  • blk.57.attn_norm.weight
    F32
    [4096]
  • blk.57.ssm_a
    F32
    [1, 128]
  • blk.57.ssm_conv1d.bias
    F32
    [10240]
  • blk.57.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.57.ssm_d
    F32
    [1, 128]
  • blk.57.ssm_dt.bias
    F32
    [128]
  • blk.57.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.57.ssm_norm.weight
    F32
    [1024, 8]
  • blk.57.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.58
  • blk.58.attn_k.weight
    Q8_0
    [4096, 256]
  • blk.58.attn_norm.weight
    F32
    [4096]
  • blk.58.attn_output.weight
    Q8_0
    [4096, 4096]
  • blk.58.attn_q.weight
    Q8_0
    [4096, 4096]
  • blk.58.attn_v.weight
    Q8_0
    [4096, 256]
  • blk.59
  • blk.59.attn_norm.weight
    F32
    [4096]
  • blk.59.exp_probs_b.bias
    F32
    [512]
  • blk.59.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.59.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.59.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.59.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.59.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.59.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.59.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.60
  • blk.60.attn_norm.weight
    F32
    [4096]
  • blk.60.ssm_a
    F32
    [1, 128]
  • blk.60.ssm_conv1d.bias
    F32
    [10240]
  • blk.60.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.60.ssm_d
    F32
    [1, 128]
  • blk.60.ssm_dt.bias
    F32
    [128]
  • blk.60.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.60.ssm_norm.weight
    F32
    [1024, 8]
  • blk.60.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.61
  • blk.61.attn_norm.weight
    F32
    [4096]
  • blk.61.exp_probs_b.bias
    F32
    [512]
  • blk.61.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.61.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.61.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.61.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.61.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.61.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.61.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.62
  • blk.62.attn_norm.weight
    F32
    [4096]
  • blk.62.ssm_a
    F32
    [1, 128]
  • blk.62.ssm_conv1d.bias
    F32
    [10240]
  • blk.62.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.62.ssm_d
    F32
    [1, 128]
  • blk.62.ssm_dt.bias
    F32
    [128]
  • blk.62.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.62.ssm_norm.weight
    F32
    [1024, 8]
  • blk.62.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.63
  • blk.63.attn_norm.weight
    F32
    [4096]
  • blk.63.exp_probs_b.bias
    F32
    [512]
  • blk.63.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.63.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.63.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.63.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.63.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.63.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.63.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.64
  • blk.64.attn_norm.weight
    F32
    [4096]
  • blk.64.ssm_a
    F32
    [1, 128]
  • blk.64.ssm_conv1d.bias
    F32
    [10240]
  • blk.64.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.64.ssm_d
    F32
    [1, 128]
  • blk.64.ssm_dt.bias
    F32
    [128]
  • blk.64.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.64.ssm_norm.weight
    F32
    [1024, 8]
  • blk.64.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.65
  • blk.65.attn_norm.weight
    F32
    [4096]
  • blk.65.exp_probs_b.bias
    F32
    [512]
  • blk.65.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.65.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.65.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.65.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.65.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.65.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.65.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.66
  • blk.66.attn_norm.weight
    F32
    [4096]
  • blk.66.ssm_a
    F32
    [1, 128]
  • blk.66.ssm_conv1d.bias
    F32
    [10240]
  • blk.66.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.66.ssm_d
    F32
    [1, 128]
  • blk.66.ssm_dt.bias
    F32
    [128]
  • blk.66.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.66.ssm_norm.weight
    F32
    [1024, 8]
  • blk.66.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.67
  • blk.67.attn_norm.weight
    F32
    [4096]
  • blk.67.exp_probs_b.bias
    F32
    [512]
  • blk.67.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.67.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.67.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.67.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.67.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.67.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.67.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.68
  • blk.68.attn_norm.weight
    F32
    [4096]
  • blk.68.ssm_a
    F32
    [1, 128]
  • blk.68.ssm_conv1d.bias
    F32
    [10240]
  • blk.68.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.68.ssm_d
    F32
    [1, 128]
  • blk.68.ssm_dt.bias
    F32
    [128]
  • blk.68.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.68.ssm_norm.weight
    F32
    [1024, 8]
  • blk.68.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.69
  • blk.69.attn_k.weight
    Q8_0
    [4096, 256]
  • blk.69.attn_norm.weight
    F32
    [4096]
  • blk.69.attn_output.weight
    Q8_0
    [4096, 4096]
  • blk.69.attn_q.weight
    Q8_0
    [4096, 4096]
  • blk.69.attn_v.weight
    Q8_0
    [4096, 256]
  • blk.70
  • blk.70.attn_norm.weight
    F32
    [4096]
  • blk.70.exp_probs_b.bias
    F32
    [512]
  • blk.70.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.70.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.70.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.70.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.70.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.70.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.70.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.71
  • blk.71.attn_norm.weight
    F32
    [4096]
  • blk.71.ssm_a
    F32
    [1, 128]
  • blk.71.ssm_conv1d.bias
    F32
    [10240]
  • blk.71.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.71.ssm_d
    F32
    [1, 128]
  • blk.71.ssm_dt.bias
    F32
    [128]
  • blk.71.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.71.ssm_norm.weight
    F32
    [1024, 8]
  • blk.71.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.72
  • blk.72.attn_norm.weight
    F32
    [4096]
  • blk.72.exp_probs_b.bias
    F32
    [512]
  • blk.72.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.72.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.72.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.72.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.72.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.72.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.72.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.73
  • blk.73.attn_norm.weight
    F32
    [4096]
  • blk.73.ssm_a
    F32
    [1, 128]
  • blk.73.ssm_conv1d.bias
    F32
    [10240]
  • blk.73.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.73.ssm_d
    F32
    [1, 128]
  • blk.73.ssm_dt.bias
    F32
    [128]
  • blk.73.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.73.ssm_norm.weight
    F32
    [1024, 8]
  • blk.73.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.74
  • blk.74.attn_norm.weight
    F32
    [4096]
  • blk.74.exp_probs_b.bias
    F32
    [512]
  • blk.74.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.74.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.74.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.74.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.74.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.74.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.74.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.75
  • blk.75.attn_norm.weight
    F32
    [4096]
  • blk.75.ssm_a
    F32
    [1, 128]
  • blk.75.ssm_conv1d.bias
    F32
    [10240]
  • blk.75.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.75.ssm_d
    F32
    [1, 128]
  • blk.75.ssm_dt.bias
    F32
    [128]
  • blk.75.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.75.ssm_norm.weight
    F32
    [1024, 8]
  • blk.75.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.76
  • blk.76.attn_norm.weight
    F32
    [4096]
  • blk.76.exp_probs_b.bias
    F32
    [512]
  • blk.76.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.76.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.76.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.76.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.76.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.76.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.76.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.77
  • blk.77.attn_norm.weight
    F32
    [4096]
  • blk.77.ssm_a
    F32
    [1, 128]
  • blk.77.ssm_conv1d.bias
    F32
    [10240]
  • blk.77.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.77.ssm_d
    F32
    [1, 128]
  • blk.77.ssm_dt.bias
    F32
    [128]
  • blk.77.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.77.ssm_norm.weight
    F32
    [1024, 8]
  • blk.77.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.78
  • blk.78.attn_k.weight
    Q8_0
    [4096, 256]
  • blk.78.attn_norm.weight
    F32
    [4096]
  • blk.78.attn_output.weight
    Q8_0
    [4096, 4096]
  • blk.78.attn_q.weight
    Q8_0
    [4096, 4096]
  • blk.78.attn_v.weight
    Q8_0
    [4096, 256]
  • blk.79
  • blk.79.attn_norm.weight
    F32
    [4096]
  • blk.79.exp_probs_b.bias
    F32
    [512]
  • blk.79.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.79.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.79.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.79.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.79.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.79.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.79.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.80
  • blk.80.attn_norm.weight
    F32
    [4096]
  • blk.80.ssm_a
    F32
    [1, 128]
  • blk.80.ssm_conv1d.bias
    F32
    [10240]
  • blk.80.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.80.ssm_d
    F32
    [1, 128]
  • blk.80.ssm_dt.bias
    F32
    [128]
  • blk.80.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.80.ssm_norm.weight
    F32
    [1024, 8]
  • blk.80.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.81
  • blk.81.attn_norm.weight
    F32
    [4096]
  • blk.81.exp_probs_b.bias
    F32
    [512]
  • blk.81.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.81.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.81.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.81.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.81.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.81.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.81.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.82
  • blk.82.attn_norm.weight
    F32
    [4096]
  • blk.82.ssm_a
    F32
    [1, 128]
  • blk.82.ssm_conv1d.bias
    F32
    [10240]
  • blk.82.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.82.ssm_d
    F32
    [1, 128]
  • blk.82.ssm_dt.bias
    F32
    [128]
  • blk.82.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.82.ssm_norm.weight
    F32
    [1024, 8]
  • blk.82.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.83
  • blk.83.attn_norm.weight
    F32
    [4096]
  • blk.83.exp_probs_b.bias
    F32
    [512]
  • blk.83.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.83.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.83.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.83.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.83.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.83.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.83.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.84
  • blk.84.attn_norm.weight
    F32
    [4096]
  • blk.84.ssm_a
    F32
    [1, 128]
  • blk.84.ssm_conv1d.bias
    F32
    [10240]
  • blk.84.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.84.ssm_d
    F32
    [1, 128]
  • blk.84.ssm_dt.bias
    F32
    [128]
  • blk.84.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.84.ssm_norm.weight
    F32
    [1024, 8]
  • blk.84.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.85
  • blk.85.attn_norm.weight
    F32
    [4096]
  • blk.85.exp_probs_b.bias
    F32
    [512]
  • blk.85.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.85.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.85.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.85.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.85.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.85.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.85.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • blk.86
  • blk.86.attn_norm.weight
    F32
    [4096]
  • blk.86.ssm_a
    F32
    [1, 128]
  • blk.86.ssm_conv1d.bias
    F32
    [10240]
  • blk.86.ssm_conv1d.weight
    F32
    [4, 10240]
  • blk.86.ssm_d
    F32
    [1, 128]
  • blk.86.ssm_dt.bias
    F32
    [128]
  • blk.86.ssm_in.weight
    Q8_0
    [4096, 18560]
  • blk.86.ssm_norm.weight
    F32
    [1024, 8]
  • blk.86.ssm_out.weight
    Q8_0
    [8192, 4096]
  • blk.87
  • blk.87.attn_norm.weight
    F32
    [4096]
  • blk.87.exp_probs_b.bias
    F32
    [512]
  • blk.87.ffn_down_exps.weight
    Q8_0
    [2688, 1024, 512]
  • blk.87.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • blk.87.ffn_gate_inp.weight
    F32
    [4096, 512]
  • blk.87.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • blk.87.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • blk.87.ffn_up_exps.weight
    Q8_0
    [1024, 2688, 512]
  • blk.87.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • mtp.layers.0.attn_k.weight
    Q8_0
    [4096, 256]
  • mtp.layers.0.attn_norm.weight
    F32
    [4096]
  • mtp.layers.0.attn_output.weight
    Q8_0
    [4096, 4096]
  • mtp.layers.0.attn_q.weight
    Q8_0
    [4096, 4096]
  • mtp.layers.0.attn_v.weight
    Q8_0
    [4096, 256]
  • mtp.layers.0.eh_proj.weight
    Q8_0
    [8192, 4096]
  • mtp.layers.0.enorm.weight
    F32
    [4096]
  • mtp.layers.0.hnorm.weight
    F32
    [4096]
  • mtp.layers.1.attn_norm.weight
    F32
    [4096]
  • mtp.layers.1.exp_probs_b.bias
    F32
    [512]
  • mtp.layers.1.ffn_down_shexp.weight
    Q8_0
    [5376, 4096]
  • mtp.layers.1.ffn_gate_inp.weight
    F32
    [4096, 512]
  • mtp.layers.1.ffn_latent_in.weight
    Q8_0
    [4096, 1024]
  • mtp.layers.1.ffn_latent_out.weight
    Q8_0
    [1024, 4096]
  • mtp.layers.1.ffn_up_shexp.weight
    Q8_0
    [4096, 5376]
  • mtp.layers.1.final_layernorm.weight
    F32
    [4096]
  • mtp.layers.1.mixer.experts.0.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.0.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.1.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.1.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.2.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.2.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.3.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.3.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.4.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.4.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.5.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.5.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.6.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.6.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.7.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.7.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.8.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.8.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.9.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.9.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.10.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.10.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.11.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.11.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.12.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.12.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.13.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.13.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.14.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.14.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.15.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.15.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.16.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.16.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.17.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.17.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.18.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.18.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.19.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.19.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.20.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.20.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.21.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.21.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.22.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.22.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.23.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.23.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.24.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.24.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.25.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.25.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.26.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.26.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.27.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.27.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.28.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.28.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.29.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.29.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.30.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.30.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.31.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.31.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.32.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.32.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.33.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.33.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.34.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.34.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.35.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.35.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.36.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.36.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.37.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.37.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.38.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.38.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.39.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.39.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.40.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.40.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.41.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.41.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.42.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.42.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.43.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.43.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.44.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.44.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.45.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.45.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.46.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.46.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.47.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.47.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.48.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.48.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.49.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.49.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.50.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.50.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.51.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.51.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.52.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.52.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.53.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.53.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.54.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.54.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.55.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.55.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.56.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.56.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.57.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.57.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.58.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.58.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.59.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.59.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.60.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.60.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.61.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.61.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.62.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.62.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.63.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.63.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.64.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.64.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.65.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.65.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.66.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.66.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.67.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.67.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.68.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.68.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.69.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.69.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.70.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.70.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.71.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.71.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.72.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.72.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.73.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.73.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.74.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.74.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.75.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.75.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.76.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.76.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.77.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.77.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.78.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.78.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.79.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.79.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.80.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.80.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.81.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.81.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.82.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.82.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.83.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.83.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.84.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.84.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.85.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.85.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.86.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.86.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.87.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.87.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.88.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.88.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.89.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.89.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.90.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.90.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.91.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.91.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.92.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.92.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.93.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.93.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.94.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.94.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.95.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.95.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.96.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.96.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.97.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.97.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.98.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.98.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.99.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.99.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.100.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.100.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.101.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.101.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.102.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.102.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.103.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.103.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.104.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.104.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.105.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.105.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.106.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.106.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.107.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.107.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.108.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.108.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.109.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.109.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.110.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.110.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.111.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.111.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.112.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.112.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.113.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.113.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.114.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.114.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.115.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.115.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.116.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.116.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.117.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.117.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.118.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.118.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.119.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.119.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.120.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.120.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.121.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.121.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.122.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.122.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.123.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.123.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.124.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.124.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.125.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.125.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.126.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.126.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.127.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.127.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.128.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.128.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.129.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.129.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.130.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.130.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.131.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.131.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.132.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.132.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.133.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.133.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.134.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.134.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.135.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.135.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.136.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.136.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.137.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.137.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.138.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.138.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.139.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.139.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.140.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.140.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.141.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.141.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.142.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.142.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.143.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.143.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.144.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.144.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.145.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.145.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.146.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.146.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.147.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.147.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.148.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.148.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.149.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.149.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.150.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.150.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.151.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.151.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.152.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.152.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.153.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.153.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.154.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.154.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.155.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.155.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.156.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.156.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.157.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.157.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.158.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.158.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.159.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.159.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.160.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.160.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.161.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.161.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.162.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.162.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.163.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.163.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.164.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.164.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.165.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.165.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.166.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.166.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.167.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.167.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.168.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.168.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.169.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.169.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.170.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.170.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.171.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.171.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.172.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.172.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.173.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.173.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.174.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.174.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.175.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.175.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.176.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.176.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.177.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.177.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.178.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.178.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.179.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.179.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.180.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.180.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.181.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.181.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.182.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.182.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.183.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.183.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.184.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.184.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.185.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.185.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.186.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.186.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.187.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.187.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.188.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.188.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.189.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.189.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.190.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.190.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.191.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.191.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.192.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.192.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.193.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.193.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.194.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.194.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.195.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.195.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.196.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.196.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.197.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.197.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.198.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.198.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.199.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.199.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.200.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.200.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.201.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.201.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.202.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.202.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.203.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.203.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.204.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.204.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.205.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.205.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.206.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.206.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.207.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.207.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.208.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.208.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.209.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.209.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.210.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.210.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.211.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.211.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.212.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.212.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.213.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.213.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.214.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.214.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.215.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.215.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.216.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.216.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.217.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.217.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.218.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.218.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.219.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.219.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.220.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.220.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.221.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.221.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.222.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.222.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.223.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.223.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.224.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.224.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.225.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.225.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.226.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.226.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.227.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.227.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.228.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.228.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.229.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.229.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.230.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.230.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.231.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.231.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.232.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.232.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.233.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.233.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.234.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.234.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.235.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.235.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.236.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.236.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.237.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.237.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.238.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.238.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.239.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.239.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.240.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.240.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.241.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.241.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.242.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.242.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.243.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.243.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.244.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.244.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.245.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.245.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.246.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.246.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.247.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.247.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.248.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.248.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.249.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.249.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.250.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.250.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.251.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.251.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.252.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.252.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.253.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.253.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.254.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.254.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.255.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.255.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.256.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.256.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.257.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.257.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.258.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.258.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.259.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.259.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.260.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.260.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.261.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.261.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.262.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.262.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.263.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.263.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.264.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.264.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.265.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.265.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.266.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.266.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.267.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.267.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.268.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.268.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.269.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.269.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.270.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.270.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.271.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.271.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.272.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.272.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.273.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.273.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.274.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.274.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.275.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.275.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.276.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.276.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.277.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.277.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.278.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.278.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.279.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.279.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.280.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.280.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.281.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.281.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.282.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.282.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.283.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.283.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.284.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.284.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.285.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.285.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.286.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.286.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.287.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.287.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.288.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.288.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.289.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.289.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.290.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.290.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.291.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.291.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.292.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.292.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.293.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.293.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.294.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.294.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.295.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.295.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.296.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.296.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.297.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.297.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.298.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.298.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.299.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.299.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.300.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.300.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.301.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.301.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.302.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.302.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.303.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.303.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.304.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.304.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.305.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.305.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.306.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.306.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.307.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.307.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.308.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.308.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.309.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.309.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.310.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.310.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.311.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.311.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.312.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.312.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.313.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.313.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.314.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.314.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.315.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.315.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.316.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.316.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.317.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.317.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.318.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.318.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.319.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.319.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.320.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.320.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.321.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.321.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.322.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.322.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.323.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.323.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.324.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.324.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.325.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.325.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.326.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.326.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.327.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.327.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.328.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.328.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.329.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.329.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.330.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.330.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.331.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.331.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.332.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.332.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.333.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.333.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.334.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.334.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.335.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.335.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.336.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.336.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.337.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.337.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.338.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.338.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.339.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.339.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.340.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.340.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.341.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.341.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.342.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.342.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.343.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.343.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.344.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.344.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.345.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.345.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.346.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.346.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.347.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.347.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.348.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.348.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.349.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.349.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.350.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.350.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.351.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.351.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.352.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.352.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.353.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.353.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.354.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.354.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.355.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.355.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.356.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.356.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.357.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.357.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.358.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.358.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.359.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.359.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.360.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.360.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.361.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.361.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.362.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.362.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.363.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.363.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.364.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.364.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.365.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.365.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.366.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.366.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.367.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.367.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.368.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.368.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.369.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.369.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.370.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.370.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.371.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.371.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.372.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.372.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.373.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.373.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.374.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.374.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.375.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.375.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.376.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.376.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.377.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.377.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.378.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.378.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.379.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.379.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.380.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.380.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.381.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.381.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.382.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.382.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.383.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.383.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.384.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.384.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.385.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.385.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.386.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.386.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.387.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.387.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.388.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.388.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.389.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.389.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.390.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.390.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.391.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.391.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.392.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.392.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.393.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.393.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.394.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.394.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.395.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.395.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.396.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.396.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.397.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.397.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.398.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.398.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.399.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.399.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.400.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.400.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.401.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.401.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.402.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.402.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.403.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.403.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.404.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.404.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.405.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.405.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.406.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.406.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.407.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.407.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.408.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.408.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.409.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.409.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.410.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.410.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.411.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.411.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.412.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.412.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.413.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.413.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.414.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.414.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.415.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.415.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.416.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.416.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.417.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.417.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.418.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.418.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.419.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.419.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.420.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.420.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.421.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.421.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.422.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.422.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.423.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.423.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.424.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.424.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.425.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.425.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.426.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.426.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.427.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.427.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.428.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.428.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.429.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.429.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.430.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.430.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.431.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.431.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.432.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.432.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.433.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.433.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.434.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.434.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.435.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.435.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.436.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.436.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.437.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.437.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.438.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.438.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.439.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.439.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.440.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.440.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.441.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.441.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.442.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.442.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.443.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.443.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.444.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.444.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.445.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.445.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.446.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.446.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.447.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.447.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.448.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.448.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.449.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.449.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.450.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.450.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.451.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.451.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.452.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.452.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.453.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.453.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.454.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.454.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.455.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.455.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.456.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.456.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.457.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.457.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.458.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.458.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.459.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.459.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.460.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.460.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.461.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.461.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.462.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.462.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.463.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.463.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.464.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.464.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.465.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.465.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.466.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.466.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.467.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.467.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.468.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.468.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.469.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.469.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.470.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.470.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.471.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.471.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.472.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.472.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.473.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.473.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.474.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.474.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.475.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.475.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.476.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.476.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.477.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.477.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.478.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.478.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.479.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.479.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.480.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.480.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.481.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.481.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.482.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.482.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.483.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.483.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.484.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.484.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.485.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.485.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.486.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.486.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.487.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.487.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.488.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.488.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.489.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.489.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.490.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.490.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.491.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.491.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.492.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.492.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.493.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.493.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.494.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.494.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.495.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.495.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.496.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.496.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.497.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.497.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.498.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.498.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.499.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.499.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.500.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.500.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.501.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.501.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.502.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.502.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.503.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.503.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.504.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.504.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.505.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.505.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.506.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.506.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.507.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.507.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.508.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.508.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.509.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.509.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.510.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.510.up_proj.weight
    Q8_0
    [1024, 2688]
  • mtp.layers.1.mixer.experts.511.down_proj.weight
    Q8_0
    [2688, 1024]
  • mtp.layers.1.mixer.experts.511.up_proj.weight
    Q8_0
    [1024, 2688]
  • output.weight
    Q8_0
    [4096, 131072]
  • output_norm.weight
    F32
    [4096]