latest
4.9GB
[LLAMA TOOL TEMPLATE] Llama 3.1 is a new state-of-the-art model from Meta available with tool use.
Tools
8B
128 Pulls Updated 8 weeks ago
8e7d07fdf167 · 4.9GB
-
general.architecturellama
-
general.file_typeQ4_K_M
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length8192
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base500000
-
llama.vocab_size128256
-
tokenizer.ggml.bos_token_id128000
-
tokenizer.ggml.eos_token_id128001
-
tokenizer.ggml.merges[Ġ Ġ Ġ ĠĠĠ ĠĠ ĠĠ ĠĠĠ Ġ i n ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.presmaug-bpe
-
tokenizer.ggml.token_type[1 1 1 1 1 ...]
-
tokenizer.ggml.tokens[! " # $ % ...]
-
NameTypeShape
-
token_embd.weightQ4_K[4096 128256]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightQ6_K[14336 4096]
-
blk.0.ffn_gate.weightQ4_K[4096 14336]
-
blk.0.ffn_up.weightQ4_K[4096 14336]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightQ4_K[4096 1024]
-
blk.0.attn_output.weightQ4_K[4096 4096]
-
blk.0.attn_q.weightQ4_K[4096 4096]
-
blk.0.attn_v.weightQ6_K[4096 1024]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightQ6_K[14336 4096]
-
blk.1.ffn_gate.weightQ4_K[4096 14336]
-
blk.1.ffn_up.weightQ4_K[4096 14336]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightQ4_K[4096 1024]
-
blk.1.attn_output.weightQ4_K[4096 4096]
-
blk.1.attn_q.weightQ4_K[4096 4096]
-
blk.1.attn_v.weightQ6_K[4096 1024]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightQ6_K[14336 4096]
-
blk.2.ffn_gate.weightQ4_K[4096 14336]
-
blk.2.ffn_up.weightQ4_K[4096 14336]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightQ4_K[4096 1024]
-
blk.2.attn_output.weightQ4_K[4096 4096]
-
blk.2.attn_q.weightQ4_K[4096 4096]
-
blk.2.attn_v.weightQ6_K[4096 1024]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightQ6_K[14336 4096]
-
blk.3.ffn_gate.weightQ4_K[4096 14336]
-
blk.3.ffn_up.weightQ4_K[4096 14336]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightQ4_K[4096 1024]
-
blk.3.attn_output.weightQ4_K[4096 4096]
-
blk.3.attn_q.weightQ4_K[4096 4096]
-
blk.3.attn_v.weightQ6_K[4096 1024]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightQ4_K[14336 4096]
-
blk.4.ffn_gate.weightQ4_K[4096 14336]
-
blk.4.ffn_up.weightQ4_K[4096 14336]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightQ4_K[4096 1024]
-
blk.4.attn_output.weightQ4_K[4096 4096]
-
blk.4.attn_q.weightQ4_K[4096 4096]
-
blk.4.attn_v.weightQ4_K[4096 1024]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightQ4_K[14336 4096]
-
blk.5.ffn_gate.weightQ4_K[4096 14336]
-
blk.5.ffn_up.weightQ4_K[4096 14336]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightQ4_K[4096 1024]
-
blk.5.attn_output.weightQ4_K[4096 4096]
-
blk.5.attn_q.weightQ4_K[4096 4096]
-
blk.5.attn_v.weightQ4_K[4096 1024]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightQ6_K[14336 4096]
-
blk.6.ffn_gate.weightQ4_K[4096 14336]
-
blk.6.ffn_up.weightQ4_K[4096 14336]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightQ4_K[4096 1024]
-
blk.6.attn_output.weightQ4_K[4096 4096]
-
blk.6.attn_q.weightQ4_K[4096 4096]
-
blk.6.attn_v.weightQ6_K[4096 1024]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightQ4_K[14336 4096]
-
blk.7.ffn_gate.weightQ4_K[4096 14336]
-
blk.7.ffn_up.weightQ4_K[4096 14336]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightQ4_K[4096 1024]
-
blk.7.attn_output.weightQ4_K[4096 4096]
-
blk.7.attn_q.weightQ4_K[4096 4096]
-
blk.7.attn_v.weightQ4_K[4096 1024]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightQ4_K[14336 4096]
-
blk.8.ffn_gate.weightQ4_K[4096 14336]
-
blk.8.ffn_up.weightQ4_K[4096 14336]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightQ4_K[4096 1024]
-
blk.8.attn_output.weightQ4_K[4096 4096]
-
blk.8.attn_q.weightQ4_K[4096 4096]
-
blk.8.attn_v.weightQ4_K[4096 1024]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightQ4_K[14336 4096]
-
blk.9.ffn_gate.weightQ4_K[4096 14336]
-
blk.9.ffn_up.weightQ4_K[4096 14336]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightQ4_K[4096 1024]
-
blk.9.attn_output.weightQ4_K[4096 4096]
-
blk.9.attn_q.weightQ4_K[4096 4096]
-
blk.9.attn_v.weightQ4_K[4096 1024]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightQ6_K[14336 4096]
-
blk.10.ffn_gate.weightQ4_K[4096 14336]
-
blk.10.ffn_up.weightQ4_K[4096 14336]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.weightQ4_K[4096 1024]
-
blk.10.attn_output.weightQ4_K[4096 4096]
-
blk.10.attn_q.weightQ4_K[4096 4096]
-
blk.10.attn_v.weightQ6_K[4096 1024]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightQ4_K[14336 4096]
-
blk.11.ffn_gate.weightQ4_K[4096 14336]
-
blk.11.ffn_up.weightQ4_K[4096 14336]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightQ4_K[4096 1024]
-
blk.11.attn_output.weightQ4_K[4096 4096]
-
blk.11.attn_q.weightQ4_K[4096 4096]
-
blk.11.attn_v.weightQ4_K[4096 1024]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightQ4_K[14336 4096]
-
blk.12.ffn_gate.weightQ4_K[4096 14336]
-
blk.12.ffn_up.weightQ4_K[4096 14336]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightQ4_K[4096 1024]
-
blk.12.attn_output.weightQ4_K[4096 4096]
-
blk.12.attn_q.weightQ4_K[4096 4096]
-
blk.12.attn_v.weightQ4_K[4096 1024]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightQ6_K[14336 4096]
-
blk.13.ffn_gate.weightQ4_K[4096 14336]
-
blk.13.ffn_up.weightQ4_K[4096 14336]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightQ4_K[4096 1024]
-
blk.13.attn_output.weightQ4_K[4096 4096]
-
blk.13.attn_q.weightQ4_K[4096 4096]
-
blk.13.attn_v.weightQ6_K[4096 1024]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightQ4_K[14336 4096]
-
blk.14.ffn_gate.weightQ4_K[4096 14336]
-
blk.14.ffn_up.weightQ4_K[4096 14336]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightQ4_K[4096 1024]
-
blk.14.attn_output.weightQ4_K[4096 4096]
-
blk.14.attn_q.weightQ4_K[4096 4096]
-
blk.14.attn_v.weightQ4_K[4096 1024]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightQ4_K[14336 4096]
-
blk.15.ffn_gate.weightQ4_K[4096 14336]
-
blk.15.ffn_up.weightQ4_K[4096 14336]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightQ4_K[4096 1024]
-
blk.15.attn_output.weightQ4_K[4096 4096]
-
blk.15.attn_q.weightQ4_K[4096 4096]
-
blk.15.attn_v.weightQ4_K[4096 1024]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightQ6_K[14336 4096]
-
blk.16.ffn_gate.weightQ4_K[4096 14336]
-
blk.16.ffn_up.weightQ4_K[4096 14336]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightQ4_K[4096 1024]
-
blk.16.attn_output.weightQ4_K[4096 4096]
-
blk.16.attn_q.weightQ4_K[4096 4096]
-
blk.16.attn_v.weightQ6_K[4096 1024]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightQ4_K[14336 4096]
-
blk.17.ffn_gate.weightQ4_K[4096 14336]
-
blk.17.ffn_up.weightQ4_K[4096 14336]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightQ4_K[4096 1024]
-
blk.17.attn_output.weightQ4_K[4096 4096]
-
blk.17.attn_q.weightQ4_K[4096 4096]
-
blk.17.attn_v.weightQ4_K[4096 1024]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightQ4_K[14336 4096]
-
blk.18.ffn_gate.weightQ4_K[4096 14336]
-
blk.18.ffn_up.weightQ4_K[4096 14336]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightQ4_K[4096 1024]
-
blk.18.attn_output.weightQ4_K[4096 4096]
-
blk.18.attn_q.weightQ4_K[4096 4096]
-
blk.18.attn_v.weightQ4_K[4096 1024]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightQ6_K[14336 4096]
-
blk.19.ffn_gate.weightQ4_K[4096 14336]
-
blk.19.ffn_up.weightQ4_K[4096 14336]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightQ4_K[4096 1024]
-
blk.19.attn_output.weightQ4_K[4096 4096]
-
blk.19.attn_q.weightQ4_K[4096 4096]
-
blk.19.attn_v.weightQ6_K[4096 1024]
-
blk.20.ffn_gate.weightQ4_K[4096 14336]
-
blk.20.attn_k.weightQ4_K[4096 1024]
-
blk.20.attn_output.weightQ4_K[4096 4096]
-
blk.20.attn_q.weightQ4_K[4096 4096]
-
blk.20.attn_v.weightQ4_K[4096 1024]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightQ4_K[14336 4096]
-
blk.20.ffn_up.weightQ4_K[4096 14336]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightQ6_K[14336 4096]
-
blk.21.ffn_gate.weightQ4_K[4096 14336]
-
blk.21.ffn_up.weightQ4_K[4096 14336]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightQ4_K[4096 1024]
-
blk.21.attn_output.weightQ4_K[4096 4096]
-
blk.21.attn_q.weightQ4_K[4096 4096]
-
blk.21.attn_v.weightQ6_K[4096 1024]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightQ4_K[14336 4096]
-
blk.22.ffn_gate.weightQ4_K[4096 14336]
-
blk.22.ffn_up.weightQ4_K[4096 14336]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.weightQ4_K[4096 1024]
-
blk.22.attn_output.weightQ4_K[4096 4096]
-
blk.22.attn_q.weightQ4_K[4096 4096]
-
blk.22.attn_v.weightQ4_K[4096 1024]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightQ4_K[14336 4096]
-
blk.23.ffn_gate.weightQ4_K[4096 14336]
-
blk.23.ffn_up.weightQ4_K[4096 14336]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightQ4_K[4096 1024]
-
blk.23.attn_output.weightQ4_K[4096 4096]
-
blk.23.attn_q.weightQ4_K[4096 4096]
-
blk.23.attn_v.weightQ4_K[4096 1024]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightQ6_K[14336 4096]
-
blk.24.ffn_gate.weightQ4_K[4096 14336]
-
blk.24.ffn_up.weightQ4_K[4096 14336]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightQ4_K[4096 1024]
-
blk.24.attn_output.weightQ4_K[4096 4096]
-
blk.24.attn_q.weightQ4_K[4096 4096]
-
blk.24.attn_v.weightQ6_K[4096 1024]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightQ4_K[14336 4096]
-
blk.25.ffn_gate.weightQ4_K[4096 14336]
-
blk.25.ffn_up.weightQ4_K[4096 14336]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightQ4_K[4096 1024]
-
blk.25.attn_output.weightQ4_K[4096 4096]
-
blk.25.attn_q.weightQ4_K[4096 4096]
-
blk.25.attn_v.weightQ4_K[4096 1024]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightQ4_K[14336 4096]
-
blk.26.ffn_gate.weightQ4_K[4096 14336]
-
blk.26.ffn_up.weightQ4_K[4096 14336]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightQ4_K[4096 1024]
-
blk.26.attn_output.weightQ4_K[4096 4096]
-
blk.26.attn_q.weightQ4_K[4096 4096]
-
blk.26.attn_v.weightQ4_K[4096 1024]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightQ6_K[14336 4096]
-
blk.27.ffn_gate.weightQ4_K[4096 14336]
-
blk.27.ffn_up.weightQ4_K[4096 14336]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightQ4_K[4096 1024]
-
blk.27.attn_output.weightQ4_K[4096 4096]
-
blk.27.attn_q.weightQ4_K[4096 4096]
-
blk.27.attn_v.weightQ6_K[4096 1024]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightQ6_K[14336 4096]
-
blk.28.ffn_gate.weightQ4_K[4096 14336]
-
blk.28.ffn_up.weightQ4_K[4096 14336]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightQ4_K[4096 1024]
-
blk.28.attn_output.weightQ4_K[4096 4096]
-
blk.28.attn_q.weightQ4_K[4096 4096]
-
blk.28.attn_v.weightQ6_K[4096 1024]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightQ6_K[14336 4096]
-
blk.29.ffn_gate.weightQ4_K[4096 14336]
-
blk.29.ffn_up.weightQ4_K[4096 14336]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightQ4_K[4096 1024]
-
blk.29.attn_output.weightQ4_K[4096 4096]
-
blk.29.attn_q.weightQ4_K[4096 4096]
-
blk.29.attn_v.weightQ6_K[4096 1024]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightQ6_K[14336 4096]
-
blk.30.ffn_gate.weightQ4_K[4096 14336]
-
blk.30.ffn_up.weightQ4_K[4096 14336]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightQ4_K[4096 1024]
-
blk.30.attn_output.weightQ4_K[4096 4096]
-
blk.30.attn_q.weightQ4_K[4096 4096]
-
blk.30.attn_v.weightQ6_K[4096 1024]
-
blk.31.ffn_gate.weightQ4_K[4096 14336]
-
blk.31.ffn_up.weightQ4_K[4096 14336]
-
blk.31.attn_k.weightQ4_K[4096 1024]
-
blk.31.attn_output.weightQ4_K[4096 4096]
-
blk.31.attn_q.weightQ4_K[4096 4096]
-
blk.31.attn_v.weightQ6_K[4096 1024]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightQ6_K[14336 4096]
-
blk.31.ffn_norm.weightF32[4096]
-
output.weightQ6_K[4096 128256]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31