latest
14GB
7B
7 Pulls Updated 5 months ago
cbd9455acec4 · 14GB
-
general.architecturellama
-
general.file_typeF16
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length32768
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base10000
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id2
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 6 6 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
tokenizer.ggml.unknown_token_id0
-
NameTypeShape
-
token_embd.weightF16[4096 32000]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightF16[14336 4096]
-
blk.0.ffn_gate.weightF16[4096 14336]
-
blk.0.ffn_up.weightF16[4096 14336]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightF16[4096 1024]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.attn_q.weightF16[4096 4096]
-
blk.0.attn_v.weightF16[4096 1024]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightF16[14336 4096]
-
blk.1.ffn_gate.weightF16[4096 14336]
-
blk.1.ffn_up.weightF16[4096 14336]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightF16[4096 1024]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.attn_q.weightF16[4096 4096]
-
blk.1.attn_v.weightF16[4096 1024]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightF16[14336 4096]
-
blk.2.ffn_gate.weightF16[4096 14336]
-
blk.2.ffn_up.weightF16[4096 14336]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightF16[4096 1024]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.attn_q.weightF16[4096 4096]
-
blk.2.attn_v.weightF16[4096 1024]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightF16[14336 4096]
-
blk.3.ffn_gate.weightF16[4096 14336]
-
blk.3.ffn_up.weightF16[4096 14336]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightF16[4096 1024]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.attn_q.weightF16[4096 4096]
-
blk.3.attn_v.weightF16[4096 1024]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightF16[14336 4096]
-
blk.4.ffn_gate.weightF16[4096 14336]
-
blk.4.ffn_up.weightF16[4096 14336]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightF16[4096 1024]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.attn_q.weightF16[4096 4096]
-
blk.4.attn_v.weightF16[4096 1024]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightF16[14336 4096]
-
blk.5.ffn_gate.weightF16[4096 14336]
-
blk.5.ffn_up.weightF16[4096 14336]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightF16[4096 1024]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.5.attn_q.weightF16[4096 4096]
-
blk.5.attn_v.weightF16[4096 1024]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightF16[14336 4096]
-
blk.6.ffn_gate.weightF16[4096 14336]
-
blk.6.ffn_up.weightF16[4096 14336]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightF16[4096 1024]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.attn_q.weightF16[4096 4096]
-
blk.6.attn_v.weightF16[4096 1024]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightF16[14336 4096]
-
blk.7.ffn_gate.weightF16[4096 14336]
-
blk.7.ffn_up.weightF16[4096 14336]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightF16[4096 1024]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.attn_q.weightF16[4096 4096]
-
blk.7.attn_v.weightF16[4096 1024]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightF16[14336 4096]
-
blk.8.ffn_gate.weightF16[4096 14336]
-
blk.8.ffn_up.weightF16[4096 14336]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightF16[4096 1024]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.attn_q.weightF16[4096 4096]
-
blk.8.attn_v.weightF16[4096 1024]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightF16[14336 4096]
-
blk.9.ffn_gate.weightF16[4096 14336]
-
blk.9.ffn_up.weightF16[4096 14336]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightF16[4096 1024]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.attn_q.weightF16[4096 4096]
-
blk.9.attn_v.weightF16[4096 1024]
-
blk.10.ffn_gate.weightF16[4096 14336]
-
blk.10.ffn_up.weightF16[4096 14336]
-
blk.10.attn_k.weightF16[4096 1024]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.10.attn_q.weightF16[4096 4096]
-
blk.10.attn_v.weightF16[4096 1024]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightF16[14336 4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightF16[14336 4096]
-
blk.11.ffn_gate.weightF16[4096 14336]
-
blk.11.ffn_up.weightF16[4096 14336]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightF16[4096 1024]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.attn_q.weightF16[4096 4096]
-
blk.11.attn_v.weightF16[4096 1024]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightF16[14336 4096]
-
blk.12.ffn_gate.weightF16[4096 14336]
-
blk.12.ffn_up.weightF16[4096 14336]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightF16[4096 1024]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.attn_q.weightF16[4096 4096]
-
blk.12.attn_v.weightF16[4096 1024]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightF16[14336 4096]
-
blk.13.ffn_gate.weightF16[4096 14336]
-
blk.13.ffn_up.weightF16[4096 14336]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightF16[4096 1024]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.attn_q.weightF16[4096 4096]
-
blk.13.attn_v.weightF16[4096 1024]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightF16[14336 4096]
-
blk.14.ffn_gate.weightF16[4096 14336]
-
blk.14.ffn_up.weightF16[4096 14336]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightF16[4096 1024]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.attn_q.weightF16[4096 4096]
-
blk.14.attn_v.weightF16[4096 1024]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightF16[14336 4096]
-
blk.15.ffn_gate.weightF16[4096 14336]
-
blk.15.ffn_up.weightF16[4096 14336]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightF16[4096 1024]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.attn_q.weightF16[4096 4096]
-
blk.15.attn_v.weightF16[4096 1024]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightF16[14336 4096]
-
blk.16.ffn_gate.weightF16[4096 14336]
-
blk.16.ffn_up.weightF16[4096 14336]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightF16[4096 1024]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.attn_q.weightF16[4096 4096]
-
blk.16.attn_v.weightF16[4096 1024]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightF16[14336 4096]
-
blk.17.ffn_gate.weightF16[4096 14336]
-
blk.17.ffn_up.weightF16[4096 14336]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightF16[4096 1024]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.attn_q.weightF16[4096 4096]
-
blk.17.attn_v.weightF16[4096 1024]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightF16[14336 4096]
-
blk.18.ffn_gate.weightF16[4096 14336]
-
blk.18.ffn_up.weightF16[4096 14336]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightF16[4096 1024]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.attn_q.weightF16[4096 4096]
-
blk.18.attn_v.weightF16[4096 1024]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightF16[14336 4096]
-
blk.19.ffn_gate.weightF16[4096 14336]
-
blk.19.ffn_up.weightF16[4096 14336]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightF16[4096 1024]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.attn_q.weightF16[4096 4096]
-
blk.19.attn_v.weightF16[4096 1024]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightF16[14336 4096]
-
blk.20.ffn_gate.weightF16[4096 14336]
-
blk.20.ffn_up.weightF16[4096 14336]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.weightF16[4096 1024]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.attn_q.weightF16[4096 4096]
-
blk.20.attn_v.weightF16[4096 1024]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightF16[14336 4096]
-
blk.21.ffn_gate.weightF16[4096 14336]
-
blk.21.ffn_up.weightF16[4096 14336]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightF16[4096 1024]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.attn_q.weightF16[4096 4096]
-
blk.21.attn_v.weightF16[4096 1024]
-
blk.22.attn_k.weightF16[4096 1024]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.attn_q.weightF16[4096 4096]
-
blk.22.attn_v.weightF16[4096 1024]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightF16[14336 4096]
-
blk.22.ffn_gate.weightF16[4096 14336]
-
blk.22.ffn_up.weightF16[4096 14336]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightF16[14336 4096]
-
blk.23.ffn_gate.weightF16[4096 14336]
-
blk.23.ffn_up.weightF16[4096 14336]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightF16[4096 1024]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.attn_q.weightF16[4096 4096]
-
blk.23.attn_v.weightF16[4096 1024]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightF16[14336 4096]
-
blk.24.ffn_gate.weightF16[4096 14336]
-
blk.24.ffn_up.weightF16[4096 14336]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightF16[4096 1024]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.attn_q.weightF16[4096 4096]
-
blk.24.attn_v.weightF16[4096 1024]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightF16[14336 4096]
-
blk.25.ffn_gate.weightF16[4096 14336]
-
blk.25.ffn_up.weightF16[4096 14336]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightF16[4096 1024]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.attn_q.weightF16[4096 4096]
-
blk.25.attn_v.weightF16[4096 1024]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightF16[14336 4096]
-
blk.26.ffn_gate.weightF16[4096 14336]
-
blk.26.ffn_up.weightF16[4096 14336]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightF16[4096 1024]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.attn_q.weightF16[4096 4096]
-
blk.26.attn_v.weightF16[4096 1024]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightF16[14336 4096]
-
blk.27.ffn_gate.weightF16[4096 14336]
-
blk.27.ffn_up.weightF16[4096 14336]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightF16[4096 1024]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.27.attn_q.weightF16[4096 4096]
-
blk.27.attn_v.weightF16[4096 1024]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightF16[14336 4096]
-
blk.28.ffn_gate.weightF16[4096 14336]
-
blk.28.ffn_up.weightF16[4096 14336]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightF16[4096 1024]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.attn_q.weightF16[4096 4096]
-
blk.28.attn_v.weightF16[4096 1024]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightF16[14336 4096]
-
blk.29.ffn_gate.weightF16[4096 14336]
-
blk.29.ffn_up.weightF16[4096 14336]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightF16[4096 1024]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.attn_q.weightF16[4096 4096]
-
blk.29.attn_v.weightF16[4096 1024]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightF16[14336 4096]
-
blk.30.ffn_gate.weightF16[4096 14336]
-
blk.30.ffn_up.weightF16[4096 14336]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightF16[4096 1024]
-
blk.30.attn_output.weightF16[4096 4096]
-
blk.30.attn_q.weightF16[4096 4096]
-
blk.30.attn_v.weightF16[4096 1024]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightF16[14336 4096]
-
blk.31.ffn_gate.weightF16[4096 14336]
-
blk.31.ffn_up.weightF16[4096 14336]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.weightF16[4096 1024]
-
blk.31.attn_output.weightF16[4096 4096]
-
blk.31.attn_q.weightF16[4096 4096]
-
blk.31.attn_v.weightF16[4096 1024]
-
output.weightF16[4096 32000]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31