954 6 months ago

ALLaM is a series of powerful language models designed to advance Arabic Language Technology (ALT) developed by the National Center for Artificial Intelligence (NCAI) at the Saudi Data and AI Authority (SDAIA)

7b
26f877c57743 · 4.0GB
    Metadata
  • general.architecture
    llama
  • general.file_type
    Q4_0
  • llama.attention.head_count
    32
  • llama.attention.head_count_kv
    32
  • llama.attention.layer_norm_rms_epsilon
    1e-05
  • llama.block_count
    32
  • llama.context_length
    4096
  • llama.embedding_length
    4096
  • llama.feed_forward_length
    11008
  • llama.rope.dimension_count
    128
  • llama.rope.freq_base
    10000
  • llama.vocab_size
    64000
  • tokenizer.ggml.model
    llama
  • tokenizer.ggml.pre
    default
  • tokenizer.ggml.scores
    [0, 0, 0, 0, 0, ...]
  • tokenizer.ggml.token_type
    [2, 3, 3, 1, 1, ...]
  • tokenizer.ggml.tokens
    [<unk>, <s>, </s>, 0, 1, ...]
  • Tensor
  • token_embd.weight
    Q4_0
    [4096, 64000]
  • blk.0
  • blk.0.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.0.attn_norm.weight
    F32
    [4096]
  • blk.0.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.0.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.0.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.0.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.0.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.0.ffn_norm.weight
    F32
    [4096]
  • blk.0.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.1
  • blk.1.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.1.attn_norm.weight
    F32
    [4096]
  • blk.1.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.1.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.1.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.1.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.1.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.1.ffn_norm.weight
    F32
    [4096]
  • blk.1.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.2
  • blk.2.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.2.attn_norm.weight
    F32
    [4096]
  • blk.2.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.2.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.2.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.2.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.2.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.2.ffn_norm.weight
    F32
    [4096]
  • blk.2.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.3
  • blk.3.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.3.attn_norm.weight
    F32
    [4096]
  • blk.3.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.3.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.3.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.3.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.3.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.3.ffn_norm.weight
    F32
    [4096]
  • blk.3.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.4
  • blk.4.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.4.attn_norm.weight
    F32
    [4096]
  • blk.4.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.4.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.4.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.4.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.4.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.4.ffn_norm.weight
    F32
    [4096]
  • blk.4.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.5
  • blk.5.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.5.attn_norm.weight
    F32
    [4096]
  • blk.5.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.5.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.5.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.5.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.5.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.5.ffn_norm.weight
    F32
    [4096]
  • blk.5.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.6
  • blk.6.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.6.attn_norm.weight
    F32
    [4096]
  • blk.6.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.6.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.6.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.6.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.6.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.6.ffn_norm.weight
    F32
    [4096]
  • blk.6.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.7
  • blk.7.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.7.attn_norm.weight
    F32
    [4096]
  • blk.7.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.7.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.7.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.7.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.7.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.7.ffn_norm.weight
    F32
    [4096]
  • blk.7.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.8
  • blk.8.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.8.attn_norm.weight
    F32
    [4096]
  • blk.8.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.8.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.8.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.8.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.8.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.8.ffn_norm.weight
    F32
    [4096]
  • blk.8.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.9
  • blk.9.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.9.attn_norm.weight
    F32
    [4096]
  • blk.9.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.9.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.9.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.9.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.9.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.9.ffn_norm.weight
    F32
    [4096]
  • blk.9.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.10
  • blk.10.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.10.attn_norm.weight
    F32
    [4096]
  • blk.10.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.10.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.10.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.10.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.10.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.10.ffn_norm.weight
    F32
    [4096]
  • blk.10.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.11
  • blk.11.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.11.attn_norm.weight
    F32
    [4096]
  • blk.11.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.11.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.11.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.11.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.11.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.11.ffn_norm.weight
    F32
    [4096]
  • blk.11.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.12
  • blk.12.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.12.attn_norm.weight
    F32
    [4096]
  • blk.12.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.12.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.12.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.12.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.12.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.12.ffn_norm.weight
    F32
    [4096]
  • blk.12.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.13
  • blk.13.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.13.attn_norm.weight
    F32
    [4096]
  • blk.13.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.13.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.13.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.13.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.13.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.13.ffn_norm.weight
    F32
    [4096]
  • blk.13.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.14
  • blk.14.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.14.attn_norm.weight
    F32
    [4096]
  • blk.14.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.14.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.14.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.14.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.14.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.14.ffn_norm.weight
    F32
    [4096]
  • blk.14.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.15
  • blk.15.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.15.attn_norm.weight
    F32
    [4096]
  • blk.15.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.15.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.15.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.15.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.15.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.15.ffn_norm.weight
    F32
    [4096]
  • blk.15.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.16
  • blk.16.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.16.attn_norm.weight
    F32
    [4096]
  • blk.16.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.16.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.16.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.16.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.16.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.16.ffn_norm.weight
    F32
    [4096]
  • blk.16.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.17
  • blk.17.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.17.attn_norm.weight
    F32
    [4096]
  • blk.17.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.17.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.17.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.17.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.17.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.17.ffn_norm.weight
    F32
    [4096]
  • blk.17.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.18
  • blk.18.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.18.attn_norm.weight
    F32
    [4096]
  • blk.18.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.18.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.18.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.18.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.18.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.18.ffn_norm.weight
    F32
    [4096]
  • blk.18.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.19
  • blk.19.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.19.attn_norm.weight
    F32
    [4096]
  • blk.19.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.19.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.19.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.19.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.19.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.19.ffn_norm.weight
    F32
    [4096]
  • blk.19.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.20
  • blk.20.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.20.attn_norm.weight
    F32
    [4096]
  • blk.20.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.20.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.20.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.20.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.20.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.20.ffn_norm.weight
    F32
    [4096]
  • blk.20.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.21
  • blk.21.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.21.attn_norm.weight
    F32
    [4096]
  • blk.21.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.21.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.21.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.21.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.21.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.21.ffn_norm.weight
    F32
    [4096]
  • blk.21.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.22
  • blk.22.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.22.attn_norm.weight
    F32
    [4096]
  • blk.22.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.22.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.22.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.22.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.22.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.22.ffn_norm.weight
    F32
    [4096]
  • blk.22.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.23
  • blk.23.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.23.attn_norm.weight
    F32
    [4096]
  • blk.23.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.23.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.23.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.23.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.23.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.23.ffn_norm.weight
    F32
    [4096]
  • blk.23.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.24
  • blk.24.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.24.attn_norm.weight
    F32
    [4096]
  • blk.24.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.24.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.24.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.24.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.24.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.24.ffn_norm.weight
    F32
    [4096]
  • blk.24.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.25
  • blk.25.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.25.attn_norm.weight
    F32
    [4096]
  • blk.25.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.25.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.25.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.25.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.25.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.25.ffn_norm.weight
    F32
    [4096]
  • blk.25.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.26
  • blk.26.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.26.attn_norm.weight
    F32
    [4096]
  • blk.26.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.26.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.26.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.26.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.26.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.26.ffn_norm.weight
    F32
    [4096]
  • blk.26.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.27
  • blk.27.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.27.attn_norm.weight
    F32
    [4096]
  • blk.27.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.27.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.27.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.27.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.27.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.27.ffn_norm.weight
    F32
    [4096]
  • blk.27.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.28
  • blk.28.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.28.attn_norm.weight
    F32
    [4096]
  • blk.28.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.28.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.28.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.28.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.28.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.28.ffn_norm.weight
    F32
    [4096]
  • blk.28.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.29
  • blk.29.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.29.attn_norm.weight
    F32
    [4096]
  • blk.29.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.29.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.29.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.29.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.29.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.29.ffn_norm.weight
    F32
    [4096]
  • blk.29.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.30
  • blk.30.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.30.attn_norm.weight
    F32
    [4096]
  • blk.30.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.30.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.30.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.30.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.30.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.30.ffn_norm.weight
    F32
    [4096]
  • blk.30.ffn_up.weight
    Q4_0
    [4096, 11008]
  • blk.31
  • blk.31.attn_k.weight
    Q4_0
    [4096, 4096]
  • blk.31.attn_norm.weight
    F32
    [4096]
  • blk.31.attn_output.weight
    Q4_0
    [4096, 4096]
  • blk.31.attn_q.weight
    Q4_0
    [4096, 4096]
  • blk.31.attn_v.weight
    Q4_0
    [4096, 4096]
  • blk.31.ffn_down.weight
    Q4_0
    [11008, 4096]
  • blk.31.ffn_gate.weight
    Q4_0
    [4096, 11008]
  • blk.31.ffn_norm.weight
    F32
    [4096]
  • blk.31.ffn_up.weight
    Q4_0
    [4096, 11008]
  • output.weight
    Q6_K
    [4096, 64000]
  • output_norm.weight
    F32
    [4096]