latest
13GB
https://huggingface.co/meta-llama/LlamaGuard-7b
7B
1 Pull Updated 2 weeks ago
1ab0ec200f7c · 13GB
-
general.architecturellama
-
general.file_typeF16
-
llama.attention.head_count32
-
llama.attention.head_count_kv32
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length4096
-
llama.embedding_length4096
-
llama.feed_forward_length11008
-
llama.rope.dimension_count128
-
llama.rope.freq_base10000
-
llama.vocab_size32000
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.add_unknown_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id2
-
tokenizer.ggml.merges[▁ t e r i n ▁ a e n ...]
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.predefault
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 6 6 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
tokenizer.ggml.unknown_token_id0
-
NameTypeShape
-
token_embd.weightF16[4096 32000]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightF16[11008 4096]
-
blk.0.ffn_gate.weightF16[4096 11008]
-
blk.0.ffn_up.weightF16[4096 11008]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightF16[4096 4096]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.attn_q.weightF16[4096 4096]
-
blk.0.attn_v.weightF16[4096 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightF16[11008 4096]
-
blk.1.ffn_gate.weightF16[4096 11008]
-
blk.1.ffn_up.weightF16[4096 11008]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightF16[4096 4096]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.attn_q.weightF16[4096 4096]
-
blk.1.attn_v.weightF16[4096 4096]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.ffn_down.weightF16[11008 4096]
-
blk.2.ffn_gate.weightF16[4096 11008]
-
blk.2.ffn_up.weightF16[4096 11008]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.attn_k.weightF16[4096 4096]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.attn_q.weightF16[4096 4096]
-
blk.2.attn_v.weightF16[4096 4096]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_v.weightF16[4096 4096]
-
blk.3.attn_q.weightF16[4096 4096]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.attn_k.weightF16[4096 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightF16[11008 4096]
-
blk.3.ffn_gate.weightF16[4096 11008]
-
blk.3.ffn_up.weightF16[4096 11008]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_v.weightF16[4096 4096]
-
blk.4.attn_q.weightF16[4096 4096]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.attn_k.weightF16[4096 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightF16[11008 4096]
-
blk.4.ffn_gate.weightF16[4096 11008]
-
blk.4.ffn_up.weightF16[4096 11008]
-
blk.5.attn_q.weightF16[4096 4096]
-
blk.5.attn_v.weightF16[4096 4096]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightF16[4096 4096]
-
blk.5.ffn_up.weightF16[4096 11008]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightF16[11008 4096]
-
blk.5.ffn_gate.weightF16[4096 11008]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.6.attn_q.weightF16[4096 4096]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.ffn_up.weightF16[4096 11008]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_v.weightF16[4096 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightF16[11008 4096]
-
blk.6.ffn_gate.weightF16[4096 11008]
-
blk.6.attn_k.weightF16[4096 4096]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_v.weightF16[4096 4096]
-
blk.7.attn_q.weightF16[4096 4096]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.attn_k.weightF16[4096 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightF16[11008 4096]
-
blk.7.ffn_gate.weightF16[4096 11008]
-
blk.7.ffn_up.weightF16[4096 11008]
-
blk.8.attn_k.weightF16[4096 4096]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.attn_q.weightF16[4096 4096]
-
blk.8.attn_v.weightF16[4096 4096]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightF16[11008 4096]
-
blk.8.ffn_gate.weightF16[4096 11008]
-
blk.8.ffn_up.weightF16[4096 11008]
-
blk.9.ffn_gate.weightF16[4096 11008]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_v.weightF16[4096 4096]
-
blk.9.attn_q.weightF16[4096 4096]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightF16[11008 4096]
-
blk.9.attn_k.weightF16[4096 4096]
-
blk.9.ffn_up.weightF16[4096 11008]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.ffn_up.weightF16[4096 11008]
-
blk.10.ffn_gate.weightF16[4096 11008]
-
blk.10.ffn_down.weightF16[11008 4096]
-
blk.10.attn_k.weightF16[4096 4096]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.10.attn_q.weightF16[4096 4096]
-
blk.10.attn_v.weightF16[4096 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.attn_k.weightF16[4096 4096]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.attn_q.weightF16[4096 4096]
-
blk.11.attn_v.weightF16[4096 4096]
-
blk.11.ffn_gate.weightF16[4096 11008]
-
blk.11.ffn_down.weightF16[11008 4096]
-
blk.11.ffn_up.weightF16[4096 11008]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightF16[11008 4096]
-
blk.12.ffn_gate.weightF16[4096 11008]
-
blk.12.ffn_up.weightF16[4096 11008]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightF16[4096 4096]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.attn_q.weightF16[4096 4096]
-
blk.12.attn_v.weightF16[4096 4096]
-
blk.13.attn_k.weightF16[4096 4096]
-
blk.13.ffn_down.weightF16[11008 4096]
-
blk.13.ffn_gate.weightF16[4096 11008]
-
blk.13.ffn_up.weightF16[4096 11008]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.attn_q.weightF16[4096 4096]
-
blk.13.attn_v.weightF16[4096 4096]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.ffn_down.weightF16[11008 4096]
-
blk.14.ffn_gate.weightF16[4096 11008]
-
blk.14.ffn_up.weightF16[4096 11008]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.attn_k.weightF16[4096 4096]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.attn_q.weightF16[4096 4096]
-
blk.14.attn_v.weightF16[4096 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightF16[11008 4096]
-
blk.15.ffn_gate.weightF16[4096 11008]
-
blk.15.ffn_up.weightF16[4096 11008]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightF16[4096 4096]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.attn_q.weightF16[4096 4096]
-
blk.15.attn_v.weightF16[4096 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightF16[11008 4096]
-
blk.16.ffn_gate.weightF16[4096 11008]
-
blk.16.ffn_up.weightF16[4096 11008]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightF16[4096 4096]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.attn_q.weightF16[4096 4096]
-
blk.16.attn_v.weightF16[4096 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightF16[11008 4096]
-
blk.17.ffn_gate.weightF16[4096 11008]
-
blk.17.ffn_up.weightF16[4096 11008]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightF16[4096 4096]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.attn_q.weightF16[4096 4096]
-
blk.17.attn_v.weightF16[4096 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightF16[11008 4096]
-
blk.18.ffn_gate.weightF16[4096 11008]
-
blk.18.ffn_up.weightF16[4096 11008]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightF16[4096 4096]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.attn_q.weightF16[4096 4096]
-
blk.18.attn_v.weightF16[4096 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightF16[11008 4096]
-
blk.19.ffn_gate.weightF16[4096 11008]
-
blk.19.ffn_up.weightF16[4096 11008]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightF16[4096 4096]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.attn_q.weightF16[4096 4096]
-
blk.19.attn_v.weightF16[4096 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightF16[11008 4096]
-
blk.20.ffn_gate.weightF16[4096 11008]
-
blk.20.ffn_up.weightF16[4096 11008]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.weightF16[4096 4096]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.attn_q.weightF16[4096 4096]
-
blk.20.attn_v.weightF16[4096 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightF16[11008 4096]
-
blk.21.ffn_gate.weightF16[4096 11008]
-
blk.21.ffn_up.weightF16[4096 11008]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightF16[4096 4096]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.attn_q.weightF16[4096 4096]
-
blk.21.attn_v.weightF16[4096 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightF16[11008 4096]
-
blk.22.ffn_gate.weightF16[4096 11008]
-
blk.22.ffn_up.weightF16[4096 11008]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.weightF16[4096 4096]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.attn_q.weightF16[4096 4096]
-
blk.22.attn_v.weightF16[4096 4096]
-
blk.23.ffn_gate.weightF16[4096 11008]
-
blk.23.ffn_up.weightF16[4096 11008]
-
blk.23.attn_k.weightF16[4096 4096]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.attn_q.weightF16[4096 4096]
-
blk.23.attn_v.weightF16[4096 4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightF16[11008 4096]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightF16[11008 4096]
-
blk.24.ffn_gate.weightF16[4096 11008]
-
blk.24.ffn_up.weightF16[4096 11008]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightF16[4096 4096]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.attn_q.weightF16[4096 4096]
-
blk.24.attn_v.weightF16[4096 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightF16[11008 4096]
-
blk.25.ffn_gate.weightF16[4096 11008]
-
blk.25.ffn_up.weightF16[4096 11008]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightF16[4096 4096]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.attn_q.weightF16[4096 4096]
-
blk.25.attn_v.weightF16[4096 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightF16[11008 4096]
-
blk.26.ffn_gate.weightF16[4096 11008]
-
blk.26.ffn_up.weightF16[4096 11008]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightF16[4096 4096]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.attn_q.weightF16[4096 4096]
-
blk.26.attn_v.weightF16[4096 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightF16[11008 4096]
-
blk.27.ffn_gate.weightF16[4096 11008]
-
blk.27.ffn_up.weightF16[4096 11008]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightF16[4096 4096]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.27.attn_q.weightF16[4096 4096]
-
blk.27.attn_v.weightF16[4096 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightF16[11008 4096]
-
blk.28.ffn_gate.weightF16[4096 11008]
-
blk.28.ffn_up.weightF16[4096 11008]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightF16[4096 4096]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.attn_q.weightF16[4096 4096]
-
blk.28.attn_v.weightF16[4096 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightF16[11008 4096]
-
blk.29.ffn_gate.weightF16[4096 11008]
-
blk.29.ffn_up.weightF16[4096 11008]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightF16[4096 4096]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.attn_q.weightF16[4096 4096]
-
blk.29.attn_v.weightF16[4096 4096]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightF16[11008 4096]
-
blk.30.ffn_gate.weightF16[4096 11008]
-
blk.30.ffn_up.weightF16[4096 11008]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightF16[4096 4096]
-
blk.30.attn_output.weightF16[4096 4096]
-
blk.30.attn_q.weightF16[4096 4096]
-
blk.30.attn_v.weightF16[4096 4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightF16[11008 4096]
-
blk.31.ffn_gate.weightF16[4096 11008]
-
blk.31.ffn_up.weightF16[4096 11008]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.weightF16[4096 4096]
-
blk.31.attn_output.weightF16[4096 4096]
-
blk.31.attn_q.weightF16[4096 4096]
-
blk.31.attn_v.weightF16[4096 4096]
-
output.weightF16[4096 32000]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31