Merged AI model for Proton-eos with DirectX API translation, Vulkan and DirectX shader optimization, and game performance tuning. The AI will evolve and learn how to make DirectX run natively on Linux with the provided tools in the Proton install.

tools

8 4 weeks ago

6bd48eb2b42c · 822MB
    Metadata
  • general.architecture
    granitemoe
  • general.file_type
    Q4_K_M
  • granitemoe.attention.head_count
    16
  • granitemoe.attention.head_count_kv
    8
  • granitemoe.attention.layer_norm_rms_epsilon
    1e-06
  • granitemoe.attention.scale
    0.015625
  • granitemoe.block_count
    24
  • granitemoe.context_length
    4096
  • granitemoe.embedding_length
    1024
  • granitemoe.embedding_scale
    12
  • granitemoe.expert_count
    32
  • granitemoe.expert_used_count
    8
  • granitemoe.feed_forward_length
    512
  • granitemoe.logit_scale
    6
  • granitemoe.residual_scale
    0.22
  • granitemoe.rope.dimension_count
    64
  • granitemoe.rope.freq_base
    10000
  • granitemoe.vocab_size
    49155
  • tokenizer.ggml.add_bos_token
    false
  • tokenizer.ggml.add_space_prefix
    false
  • tokenizer.ggml.bos_token_id
    0
  • tokenizer.ggml.eos_token_id
    0
  • tokenizer.ggml.merges
    [Ġ Ġ, ĠĠ ĠĠ, ĠĠĠĠ ĠĠĠĠ, ĠĠ Ġ, e r, ...]
  • tokenizer.ggml.model
    gpt2
  • tokenizer.ggml.padding_token_id
    0
  • tokenizer.ggml.pre
    refact
  • tokenizer.ggml.token_type
    [3, 3, 3, 3, 3, ...]
  • tokenizer.ggml.tokens
    [<|end_of_text|>, <fim_prefix>, <fim_middle>, <fim_suffix>, <fim_pad>, ...]
  • Tensor
  • token_embd.weight
    Q6_K
    [1024, 49155]
  • blk.0
  • blk.0.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.0.attn_norm.weight
    F32
    [1024]
  • blk.0.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.0.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.0.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.0.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.0.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.0.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.0.ffn_norm.weight
    F32
    [1024]
  • blk.0.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.1
  • blk.1.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.1.attn_norm.weight
    F32
    [1024]
  • blk.1.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.1.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.1.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.1.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.1.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.1.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.1.ffn_norm.weight
    F32
    [1024]
  • blk.1.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.2
  • blk.2.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.2.attn_norm.weight
    F32
    [1024]
  • blk.2.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.2.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.2.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.2.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.2.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.2.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.2.ffn_norm.weight
    F32
    [1024]
  • blk.2.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.3
  • blk.3.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.3.attn_norm.weight
    F32
    [1024]
  • blk.3.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.3.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.3.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.3.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.3.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.3.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.3.ffn_norm.weight
    F32
    [1024]
  • blk.3.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.4
  • blk.4.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.4.attn_norm.weight
    F32
    [1024]
  • blk.4.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.4.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.4.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.4.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.4.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.4.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.4.ffn_norm.weight
    F32
    [1024]
  • blk.4.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.5
  • blk.5.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.5.attn_norm.weight
    F32
    [1024]
  • blk.5.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.5.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.5.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.5.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.5.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.5.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.5.ffn_norm.weight
    F32
    [1024]
  • blk.5.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.6
  • blk.6.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.6.attn_norm.weight
    F32
    [1024]
  • blk.6.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.6.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.6.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.6.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.6.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.6.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.6.ffn_norm.weight
    F32
    [1024]
  • blk.6.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.7
  • blk.7.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.7.attn_norm.weight
    F32
    [1024]
  • blk.7.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.7.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.7.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.7.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.7.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.7.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.7.ffn_norm.weight
    F32
    [1024]
  • blk.7.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.8
  • blk.8.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.8.attn_norm.weight
    F32
    [1024]
  • blk.8.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.8.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.8.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.8.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.8.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.8.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.8.ffn_norm.weight
    F32
    [1024]
  • blk.8.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.9
  • blk.9.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.9.attn_norm.weight
    F32
    [1024]
  • blk.9.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.9.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.9.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.9.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.9.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.9.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.9.ffn_norm.weight
    F32
    [1024]
  • blk.9.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.10
  • blk.10.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.10.attn_norm.weight
    F32
    [1024]
  • blk.10.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.10.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.10.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.10.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.10.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.10.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.10.ffn_norm.weight
    F32
    [1024]
  • blk.10.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.11
  • blk.11.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.11.attn_norm.weight
    F32
    [1024]
  • blk.11.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.11.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.11.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.11.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.11.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.11.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.11.ffn_norm.weight
    F32
    [1024]
  • blk.11.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.12
  • blk.12.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.12.attn_norm.weight
    F32
    [1024]
  • blk.12.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.12.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.12.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.12.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.12.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.12.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.12.ffn_norm.weight
    F32
    [1024]
  • blk.12.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.13
  • blk.13.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.13.attn_norm.weight
    F32
    [1024]
  • blk.13.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.13.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.13.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.13.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.13.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.13.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.13.ffn_norm.weight
    F32
    [1024]
  • blk.13.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.14
  • blk.14.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.14.attn_norm.weight
    F32
    [1024]
  • blk.14.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.14.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.14.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.14.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.14.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.14.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.14.ffn_norm.weight
    F32
    [1024]
  • blk.14.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.15
  • blk.15.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.15.attn_norm.weight
    F32
    [1024]
  • blk.15.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.15.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.15.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.15.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.15.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.15.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.15.ffn_norm.weight
    F32
    [1024]
  • blk.15.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.16
  • blk.16.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.16.attn_norm.weight
    F32
    [1024]
  • blk.16.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.16.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.16.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.16.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.16.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.16.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.16.ffn_norm.weight
    F32
    [1024]
  • blk.16.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.17
  • blk.17.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.17.attn_norm.weight
    F32
    [1024]
  • blk.17.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.17.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.17.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.17.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.17.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.17.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.17.ffn_norm.weight
    F32
    [1024]
  • blk.17.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.18
  • blk.18.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.18.attn_norm.weight
    F32
    [1024]
  • blk.18.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.18.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.18.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.18.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.18.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.18.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.18.ffn_norm.weight
    F32
    [1024]
  • blk.18.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.19
  • blk.19.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.19.attn_norm.weight
    F32
    [1024]
  • blk.19.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.19.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.19.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.19.ffn_down_exps.weight
    Q4_K
    [512, 1024, 32]
  • blk.19.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.19.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.19.ffn_norm.weight
    F32
    [1024]
  • blk.19.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.20
  • blk.20.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.20.attn_norm.weight
    F32
    [1024]
  • blk.20.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.20.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.20.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.20.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.20.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.20.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.20.ffn_norm.weight
    F32
    [1024]
  • blk.20.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.21
  • blk.21.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.21.attn_norm.weight
    F32
    [1024]
  • blk.21.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.21.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.21.attn_v.weight
    Q6_K
    [1024, 512]
  • blk.21.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.21.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.21.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.21.ffn_norm.weight
    F32
    [1024]
  • blk.21.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.22
  • blk.22.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.22.attn_norm.weight
    F32
    [1024]
  • blk.22.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.22.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.22.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.22.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.22.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.22.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.22.ffn_norm.weight
    F32
    [1024]
  • blk.22.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.23
  • blk.23.attn_k.weight
    Q4_K
    [1024, 512]
  • blk.23.attn_norm.weight
    F32
    [1024]
  • blk.23.attn_output.weight
    Q4_K
    [1024, 1024]
  • blk.23.attn_q.weight
    Q4_K
    [1024, 1024]
  • blk.23.attn_v.weight
    Q4_K
    [1024, 512]
  • blk.23.ffn_down_exps.weight
    Q6_K
    [512, 1024, 32]
  • blk.23.ffn_gate_exps.weight
    Q4_K
    [1024, 512, 32]
  • blk.23.ffn_gate_inp.weight
    F32
    [1024, 32]
  • blk.23.ffn_norm.weight
    F32
    [1024]
  • blk.23.ffn_up_exps.weight
    Q4_K
    [1024, 512, 32]
  • output_norm.weight
    F32
    [1024]