latest
14GB
7B
70 Pulls Updated 7 months ago
16a89a7f7d05 · 13GB
-
general.architecturellama
-
general.file_typeF16
-
llama.attention.head_count32
-
llama.attention.head_count_kv32
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length16384
-
llama.embedding_length4096
-
llama.feed_forward_length11008
-
llama.rope.dimension_count128
-
llama.rope.freq_base1e+06
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id2
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.padding_token_id0
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 6 6 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
tokenizer.ggml.unknown_token_id0
-
NameTypeShape
-
token_embd.weightF16[4096 32016]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightF16[11008 4096]
-
blk.0.ffn_gate.weightF16[4096 11008]
-
blk.0.ffn_up.weightF16[4096 11008]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightF16[4096 4096]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.attn_q.weightF16[4096 4096]
-
blk.0.attn_v.weightF16[4096 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightF16[11008 4096]
-
blk.1.ffn_gate.weightF16[4096 11008]
-
blk.1.ffn_up.weightF16[4096 11008]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightF16[4096 4096]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.attn_q.weightF16[4096 4096]
-
blk.1.attn_v.weightF16[4096 4096]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightF16[11008 4096]
-
blk.2.ffn_gate.weightF16[4096 11008]
-
blk.2.ffn_up.weightF16[4096 11008]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightF16[4096 4096]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.attn_q.weightF16[4096 4096]
-
blk.2.attn_v.weightF16[4096 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightF16[11008 4096]
-
blk.3.ffn_gate.weightF16[4096 11008]
-
blk.3.ffn_up.weightF16[4096 11008]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightF16[4096 4096]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.attn_q.weightF16[4096 4096]
-
blk.3.attn_v.weightF16[4096 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightF16[11008 4096]
-
blk.4.ffn_gate.weightF16[4096 11008]
-
blk.4.ffn_up.weightF16[4096 11008]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightF16[4096 4096]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.attn_q.weightF16[4096 4096]
-
blk.4.attn_v.weightF16[4096 4096]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightF16[11008 4096]
-
blk.5.ffn_gate.weightF16[4096 11008]
-
blk.5.ffn_up.weightF16[4096 11008]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightF16[4096 4096]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.5.attn_q.weightF16[4096 4096]
-
blk.5.attn_v.weightF16[4096 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightF16[11008 4096]
-
blk.6.ffn_gate.weightF16[4096 11008]
-
blk.6.ffn_up.weightF16[4096 11008]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightF16[4096 4096]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.attn_q.weightF16[4096 4096]
-
blk.6.attn_v.weightF16[4096 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightF16[11008 4096]
-
blk.7.ffn_gate.weightF16[4096 11008]
-
blk.7.ffn_up.weightF16[4096 11008]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightF16[4096 4096]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.attn_q.weightF16[4096 4096]
-
blk.7.attn_v.weightF16[4096 4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightF16[11008 4096]
-
blk.8.ffn_gate.weightF16[4096 11008]
-
blk.8.ffn_up.weightF16[4096 11008]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightF16[4096 4096]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.attn_q.weightF16[4096 4096]
-
blk.8.attn_v.weightF16[4096 4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightF16[11008 4096]
-
blk.9.ffn_gate.weightF16[4096 11008]
-
blk.9.ffn_up.weightF16[4096 11008]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightF16[4096 4096]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.attn_q.weightF16[4096 4096]
-
blk.9.attn_v.weightF16[4096 4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightF16[11008 4096]
-
blk.10.ffn_gate.weightF16[4096 11008]
-
blk.10.ffn_up.weightF16[4096 11008]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.weightF16[4096 4096]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.10.attn_q.weightF16[4096 4096]
-
blk.10.attn_v.weightF16[4096 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightF16[11008 4096]
-
blk.11.ffn_gate.weightF16[4096 11008]
-
blk.11.ffn_up.weightF16[4096 11008]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightF16[4096 4096]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.attn_q.weightF16[4096 4096]
-
blk.11.attn_v.weightF16[4096 4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightF16[11008 4096]
-
blk.12.ffn_gate.weightF16[4096 11008]
-
blk.12.ffn_up.weightF16[4096 11008]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightF16[4096 4096]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.attn_q.weightF16[4096 4096]
-
blk.12.attn_v.weightF16[4096 4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightF16[11008 4096]
-
blk.13.ffn_gate.weightF16[4096 11008]
-
blk.13.ffn_up.weightF16[4096 11008]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightF16[4096 4096]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.attn_q.weightF16[4096 4096]
-
blk.13.attn_v.weightF16[4096 4096]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightF16[11008 4096]
-
blk.14.ffn_gate.weightF16[4096 11008]
-
blk.14.ffn_up.weightF16[4096 11008]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightF16[4096 4096]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.attn_q.weightF16[4096 4096]
-
blk.14.attn_v.weightF16[4096 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightF16[11008 4096]
-
blk.15.ffn_gate.weightF16[4096 11008]
-
blk.15.ffn_up.weightF16[4096 11008]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightF16[4096 4096]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.attn_q.weightF16[4096 4096]
-
blk.15.attn_v.weightF16[4096 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightF16[11008 4096]
-
blk.16.ffn_gate.weightF16[4096 11008]
-
blk.16.ffn_up.weightF16[4096 11008]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightF16[4096 4096]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.attn_q.weightF16[4096 4096]
-
blk.16.attn_v.weightF16[4096 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightF16[11008 4096]
-
blk.17.ffn_gate.weightF16[4096 11008]
-
blk.17.ffn_up.weightF16[4096 11008]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightF16[4096 4096]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.attn_q.weightF16[4096 4096]
-
blk.17.attn_v.weightF16[4096 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightF16[11008 4096]
-
blk.18.ffn_gate.weightF16[4096 11008]
-
blk.18.ffn_up.weightF16[4096 11008]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightF16[4096 4096]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.attn_q.weightF16[4096 4096]
-
blk.18.attn_v.weightF16[4096 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightF16[11008 4096]
-
blk.19.ffn_gate.weightF16[4096 11008]
-
blk.19.ffn_up.weightF16[4096 11008]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightF16[4096 4096]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.attn_q.weightF16[4096 4096]
-
blk.19.attn_v.weightF16[4096 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightF16[11008 4096]
-
blk.20.ffn_gate.weightF16[4096 11008]
-
blk.20.ffn_up.weightF16[4096 11008]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.weightF16[4096 4096]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.attn_q.weightF16[4096 4096]
-
blk.20.attn_v.weightF16[4096 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightF16[11008 4096]
-
blk.21.ffn_gate.weightF16[4096 11008]
-
blk.21.ffn_up.weightF16[4096 11008]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightF16[4096 4096]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.attn_q.weightF16[4096 4096]
-
blk.21.attn_v.weightF16[4096 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightF16[11008 4096]
-
blk.22.ffn_gate.weightF16[4096 11008]
-
blk.22.ffn_up.weightF16[4096 11008]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.weightF16[4096 4096]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.attn_q.weightF16[4096 4096]
-
blk.22.attn_v.weightF16[4096 4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightF16[11008 4096]
-
blk.23.ffn_gate.weightF16[4096 11008]
-
blk.23.ffn_up.weightF16[4096 11008]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightF16[4096 4096]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.attn_q.weightF16[4096 4096]
-
blk.23.attn_v.weightF16[4096 4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightF16[11008 4096]
-
blk.24.ffn_gate.weightF16[4096 11008]
-
blk.24.ffn_up.weightF16[4096 11008]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightF16[4096 4096]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.attn_q.weightF16[4096 4096]
-
blk.24.attn_v.weightF16[4096 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightF16[11008 4096]
-
blk.25.ffn_gate.weightF16[4096 11008]
-
blk.25.ffn_up.weightF16[4096 11008]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightF16[4096 4096]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.attn_q.weightF16[4096 4096]
-
blk.25.attn_v.weightF16[4096 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightF16[11008 4096]
-
blk.26.ffn_gate.weightF16[4096 11008]
-
blk.26.ffn_up.weightF16[4096 11008]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightF16[4096 4096]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.attn_q.weightF16[4096 4096]
-
blk.26.attn_v.weightF16[4096 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightF16[11008 4096]
-
blk.27.ffn_gate.weightF16[4096 11008]
-
blk.27.ffn_up.weightF16[4096 11008]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightF16[4096 4096]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.27.attn_q.weightF16[4096 4096]
-
blk.27.attn_v.weightF16[4096 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightF16[11008 4096]
-
blk.28.ffn_gate.weightF16[4096 11008]
-
blk.28.ffn_up.weightF16[4096 11008]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightF16[4096 4096]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.attn_q.weightF16[4096 4096]
-
blk.28.attn_v.weightF16[4096 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightF16[11008 4096]
-
blk.29.ffn_gate.weightF16[4096 11008]
-
blk.29.ffn_up.weightF16[4096 11008]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightF16[4096 4096]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.attn_q.weightF16[4096 4096]
-
blk.29.attn_v.weightF16[4096 4096]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightF16[11008 4096]
-
blk.30.ffn_gate.weightF16[4096 11008]
-
blk.30.ffn_up.weightF16[4096 11008]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightF16[4096 4096]
-
blk.30.attn_output.weightF16[4096 4096]
-
blk.30.attn_q.weightF16[4096 4096]
-
blk.30.attn_v.weightF16[4096 4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightF16[11008 4096]
-
blk.31.ffn_gate.weightF16[4096 11008]
-
blk.31.ffn_up.weightF16[4096 11008]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.weightF16[4096 4096]
-
blk.31.attn_output.weightF16[4096 4096]
-
blk.31.attn_q.weightF16[4096 4096]
-
blk.31.attn_v.weightF16[4096 4096]
-
output.weightF16[4096 32016]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31