latest
4.8GB
xe / llamaguard
7B
100 Pulls Updated 9 months ago
7d39d2bdbdf5 · 13GB
-
general.architecturellama
-
general.file_typeF16
-
llama.attention.head_count32
-
llama.attention.head_count_kv32
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length2048
-
llama.embedding_length4096
-
llama.feed_forward_length11008
-
llama.rope.dimension_count128
-
llama.rope.freq_base10000
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id2
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 6 6 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
tokenizer.ggml.unknown_token_id0
-
NameTypeShape
-
token_embd.weightF16[4096 32000]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightF16[11008 4096]
-
blk.0.ffn_gate.weightF16[4096 11008]
-
blk.0.ffn_up.weightF16[4096 11008]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightF16[4096 4096]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.attn_q.weightF16[4096 4096]
-
blk.0.attn_v.weightF16[4096 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightF16[11008 4096]
-
blk.1.ffn_gate.weightF16[4096 11008]
-
blk.1.ffn_up.weightF16[4096 11008]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightF16[4096 4096]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.attn_q.weightF16[4096 4096]
-
blk.1.attn_v.weightF16[4096 4096]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightF16[11008 4096]
-
blk.2.ffn_gate.weightF16[4096 11008]
-
blk.2.ffn_up.weightF16[4096 11008]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightF16[4096 4096]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.attn_q.weightF16[4096 4096]
-
blk.2.attn_v.weightF16[4096 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightF16[11008 4096]
-
blk.3.ffn_gate.weightF16[4096 11008]
-
blk.3.ffn_up.weightF16[4096 11008]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightF16[4096 4096]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.attn_q.weightF16[4096 4096]
-
blk.3.attn_v.weightF16[4096 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightF16[11008 4096]
-
blk.4.ffn_gate.weightF16[4096 11008]
-
blk.4.ffn_up.weightF16[4096 11008]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightF16[4096 4096]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.attn_q.weightF16[4096 4096]
-
blk.4.attn_v.weightF16[4096 4096]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightF16[11008 4096]
-
blk.5.ffn_gate.weightF16[4096 11008]
-
blk.5.ffn_up.weightF16[4096 11008]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightF16[4096 4096]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.5.attn_q.weightF16[4096 4096]
-
blk.5.attn_v.weightF16[4096 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightF16[11008 4096]
-
blk.6.ffn_gate.weightF16[4096 11008]
-
blk.6.ffn_up.weightF16[4096 11008]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightF16[4096 4096]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.attn_q.weightF16[4096 4096]
-
blk.6.attn_v.weightF16[4096 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightF16[11008 4096]
-
blk.7.ffn_gate.weightF16[4096 11008]
-
blk.7.ffn_up.weightF16[4096 11008]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightF16[4096 4096]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.attn_q.weightF16[4096 4096]
-
blk.7.attn_v.weightF16[4096 4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightF16[11008 4096]
-
blk.8.ffn_gate.weightF16[4096 11008]
-
blk.8.ffn_up.weightF16[4096 11008]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightF16[4096 4096]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.attn_q.weightF16[4096 4096]
-
blk.8.attn_v.weightF16[4096 4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightF16[11008 4096]
-
blk.9.ffn_gate.weightF16[4096 11008]
-
blk.9.ffn_up.weightF16[4096 11008]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightF16[4096 4096]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.attn_q.weightF16[4096 4096]
-
blk.9.attn_v.weightF16[4096 4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightF16[11008 4096]
-
blk.10.ffn_gate.weightF16[4096 11008]
-
blk.10.ffn_up.weightF16[4096 11008]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.weightF16[4096 4096]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.10.attn_q.weightF16[4096 4096]
-
blk.10.attn_v.weightF16[4096 4096]
-
blk.11.ffn_gate.weightF16[4096 11008]
-
blk.11.attn_k.weightF16[4096 4096]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.attn_q.weightF16[4096 4096]
-
blk.11.attn_v.weightF16[4096 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightF16[11008 4096]
-
blk.11.ffn_up.weightF16[4096 11008]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightF16[11008 4096]
-
blk.12.ffn_gate.weightF16[4096 11008]
-
blk.12.ffn_up.weightF16[4096 11008]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightF16[4096 4096]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.attn_q.weightF16[4096 4096]
-
blk.12.attn_v.weightF16[4096 4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightF16[11008 4096]
-
blk.13.ffn_gate.weightF16[4096 11008]
-
blk.13.ffn_up.weightF16[4096 11008]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightF16[4096 4096]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.attn_q.weightF16[4096 4096]
-
blk.13.attn_v.weightF16[4096 4096]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightF16[11008 4096]
-
blk.14.ffn_gate.weightF16[4096 11008]
-
blk.14.ffn_up.weightF16[4096 11008]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightF16[4096 4096]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.attn_q.weightF16[4096 4096]
-
blk.14.attn_v.weightF16[4096 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightF16[11008 4096]
-
blk.15.ffn_gate.weightF16[4096 11008]
-
blk.15.ffn_up.weightF16[4096 11008]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightF16[4096 4096]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.attn_q.weightF16[4096 4096]
-
blk.15.attn_v.weightF16[4096 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightF16[11008 4096]
-
blk.16.ffn_gate.weightF16[4096 11008]
-
blk.16.ffn_up.weightF16[4096 11008]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightF16[4096 4096]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.attn_q.weightF16[4096 4096]
-
blk.16.attn_v.weightF16[4096 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightF16[11008 4096]
-
blk.17.ffn_gate.weightF16[4096 11008]
-
blk.17.ffn_up.weightF16[4096 11008]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightF16[4096 4096]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.attn_q.weightF16[4096 4096]
-
blk.17.attn_v.weightF16[4096 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightF16[11008 4096]
-
blk.18.ffn_gate.weightF16[4096 11008]
-
blk.18.ffn_up.weightF16[4096 11008]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightF16[4096 4096]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.attn_q.weightF16[4096 4096]
-
blk.18.attn_v.weightF16[4096 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightF16[11008 4096]
-
blk.19.ffn_gate.weightF16[4096 11008]
-
blk.19.ffn_up.weightF16[4096 11008]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightF16[4096 4096]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.attn_q.weightF16[4096 4096]
-
blk.19.attn_v.weightF16[4096 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightF16[11008 4096]
-
blk.20.ffn_gate.weightF16[4096 11008]
-
blk.20.ffn_up.weightF16[4096 11008]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.weightF16[4096 4096]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.attn_q.weightF16[4096 4096]
-
blk.20.attn_v.weightF16[4096 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightF16[11008 4096]
-
blk.21.ffn_gate.weightF16[4096 11008]
-
blk.21.ffn_up.weightF16[4096 11008]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightF16[4096 4096]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.attn_q.weightF16[4096 4096]
-
blk.21.attn_v.weightF16[4096 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightF16[11008 4096]
-
blk.22.ffn_gate.weightF16[4096 11008]
-
blk.22.ffn_up.weightF16[4096 11008]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.weightF16[4096 4096]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.attn_q.weightF16[4096 4096]
-
blk.22.attn_v.weightF16[4096 4096]
-
blk.23.ffn_gate.weightF16[4096 11008]
-
blk.23.ffn_up.weightF16[4096 11008]
-
blk.23.attn_k.weightF16[4096 4096]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.attn_q.weightF16[4096 4096]
-
blk.23.attn_v.weightF16[4096 4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightF16[11008 4096]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightF16[11008 4096]
-
blk.24.ffn_gate.weightF16[4096 11008]
-
blk.24.ffn_up.weightF16[4096 11008]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightF16[4096 4096]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.attn_q.weightF16[4096 4096]
-
blk.24.attn_v.weightF16[4096 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightF16[11008 4096]
-
blk.25.ffn_gate.weightF16[4096 11008]
-
blk.25.ffn_up.weightF16[4096 11008]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightF16[4096 4096]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.attn_q.weightF16[4096 4096]
-
blk.25.attn_v.weightF16[4096 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightF16[11008 4096]
-
blk.26.ffn_gate.weightF16[4096 11008]
-
blk.26.ffn_up.weightF16[4096 11008]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightF16[4096 4096]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.attn_q.weightF16[4096 4096]
-
blk.26.attn_v.weightF16[4096 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightF16[11008 4096]
-
blk.27.ffn_gate.weightF16[4096 11008]
-
blk.27.ffn_up.weightF16[4096 11008]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightF16[4096 4096]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.27.attn_q.weightF16[4096 4096]
-
blk.27.attn_v.weightF16[4096 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightF16[11008 4096]
-
blk.28.ffn_gate.weightF16[4096 11008]
-
blk.28.ffn_up.weightF16[4096 11008]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightF16[4096 4096]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.attn_q.weightF16[4096 4096]
-
blk.28.attn_v.weightF16[4096 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightF16[11008 4096]
-
blk.29.ffn_gate.weightF16[4096 11008]
-
blk.29.ffn_up.weightF16[4096 11008]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightF16[4096 4096]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.attn_q.weightF16[4096 4096]
-
blk.29.attn_v.weightF16[4096 4096]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightF16[11008 4096]
-
blk.30.ffn_gate.weightF16[4096 11008]
-
blk.30.ffn_up.weightF16[4096 11008]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightF16[4096 4096]
-
blk.30.attn_output.weightF16[4096 4096]
-
blk.30.attn_q.weightF16[4096 4096]
-
blk.30.attn_v.weightF16[4096 4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightF16[11008 4096]
-
blk.31.ffn_gate.weightF16[4096 11008]
-
blk.31.ffn_up.weightF16[4096 11008]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.weightF16[4096 4096]
-
blk.31.attn_output.weightF16[4096 4096]
-
blk.31.attn_q.weightF16[4096 4096]
-
blk.31.attn_v.weightF16[4096 4096]
-
output.weightF16[4096 32000]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31