latest
15GB
7B
3 Pulls Updated 2 months ago
b12448bbda4f · 15GB
-
general.architectureqwen2
-
general.file_typeF16
-
qwen2.attention.head_count32
-
qwen2.attention.head_count_kv4
-
qwen2.attention.layer_norm_rms_epsilon1e-05
-
qwen2.block_count32
-
qwen2.context_length65536
-
qwen2.embedding_length4096
-
qwen2.feed_forward_length13440
-
qwen2.rope.freq_base1e+06
-
tokenizer.ggml.add_bos_tokenfalse
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id2
-
tokenizer.ggml.eos_token_id4
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.padding_token_id92298
-
tokenizer.ggml.predefault
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 3 3 ...]
-
tokenizer.ggml.tokens[<unk> <s> <|endoftext|> <|im_start|> <|im_end|> ...]
-
tokenizer.ggml.unknown_token_id0
-
NameTypeShape
-
token_embd.weightF16[4096 92416]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightF16[13440 4096]
-
blk.0.ffn_gate.weightF16[4096 13440]
-
blk.0.ffn_up.weightF16[4096 13440]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.biasF32[512]
-
blk.0.attn_k.weightF16[4096 512]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.attn_q.biasF32[4096]
-
blk.0.attn_q.weightF16[4096 4096]
-
blk.0.attn_v.biasF32[512]
-
blk.0.attn_v.weightF16[4096 512]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightF16[13440 4096]
-
blk.1.ffn_gate.weightF16[4096 13440]
-
blk.1.ffn_up.weightF16[4096 13440]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.biasF32[512]
-
blk.1.attn_k.weightF16[4096 512]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.attn_q.biasF32[4096]
-
blk.1.attn_q.weightF16[4096 4096]
-
blk.1.attn_v.biasF32[512]
-
blk.1.attn_v.weightF16[4096 512]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightF16[13440 4096]
-
blk.2.ffn_gate.weightF16[4096 13440]
-
blk.2.ffn_up.weightF16[4096 13440]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.biasF32[512]
-
blk.2.attn_k.weightF16[4096 512]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.attn_q.biasF32[4096]
-
blk.2.attn_q.weightF16[4096 4096]
-
blk.2.attn_v.biasF32[512]
-
blk.2.attn_v.weightF16[4096 512]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightF16[13440 4096]
-
blk.3.ffn_gate.weightF16[4096 13440]
-
blk.3.ffn_up.weightF16[4096 13440]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.biasF32[512]
-
blk.3.attn_k.weightF16[4096 512]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.attn_q.biasF32[4096]
-
blk.3.attn_q.weightF16[4096 4096]
-
blk.3.attn_v.biasF32[512]
-
blk.3.attn_v.weightF16[4096 512]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightF16[13440 4096]
-
blk.4.ffn_gate.weightF16[4096 13440]
-
blk.4.ffn_up.weightF16[4096 13440]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.biasF32[512]
-
blk.4.attn_k.weightF16[4096 512]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.attn_q.biasF32[4096]
-
blk.4.attn_q.weightF16[4096 4096]
-
blk.4.attn_v.biasF32[512]
-
blk.4.attn_v.weightF16[4096 512]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightF16[13440 4096]
-
blk.5.ffn_gate.weightF16[4096 13440]
-
blk.5.ffn_up.weightF16[4096 13440]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.biasF32[512]
-
blk.5.attn_k.weightF16[4096 512]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.5.attn_q.biasF32[4096]
-
blk.5.attn_q.weightF16[4096 4096]
-
blk.5.attn_v.biasF32[512]
-
blk.5.attn_v.weightF16[4096 512]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightF16[13440 4096]
-
blk.6.ffn_gate.weightF16[4096 13440]
-
blk.6.ffn_up.weightF16[4096 13440]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.biasF32[512]
-
blk.6.attn_k.weightF16[4096 512]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.attn_q.biasF32[4096]
-
blk.6.attn_q.weightF16[4096 4096]
-
blk.6.attn_v.biasF32[512]
-
blk.6.attn_v.weightF16[4096 512]
-
blk.7.ffn_gate.weightF16[4096 13440]
-
blk.7.ffn_up.weightF16[4096 13440]
-
blk.7.attn_k.biasF32[512]
-
blk.7.attn_k.weightF16[4096 512]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.attn_q.biasF32[4096]
-
blk.7.attn_q.weightF16[4096 4096]
-
blk.7.attn_v.biasF32[512]
-
blk.7.attn_v.weightF16[4096 512]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightF16[13440 4096]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightF16[13440 4096]
-
blk.8.ffn_gate.weightF16[4096 13440]
-
blk.8.ffn_up.weightF16[4096 13440]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.biasF32[512]
-
blk.8.attn_k.weightF16[4096 512]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.attn_q.biasF32[4096]
-
blk.8.attn_q.weightF16[4096 4096]
-
blk.8.attn_v.biasF32[512]
-
blk.8.attn_v.weightF16[4096 512]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightF16[13440 4096]
-
blk.9.ffn_gate.weightF16[4096 13440]
-
blk.9.ffn_up.weightF16[4096 13440]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.biasF32[512]
-
blk.9.attn_k.weightF16[4096 512]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.attn_q.biasF32[4096]
-
blk.9.attn_q.weightF16[4096 4096]
-
blk.9.attn_v.biasF32[512]
-
blk.9.attn_v.weightF16[4096 512]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightF16[13440 4096]
-
blk.10.ffn_gate.weightF16[4096 13440]
-
blk.10.ffn_up.weightF16[4096 13440]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.biasF32[512]
-
blk.10.attn_k.weightF16[4096 512]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.10.attn_q.biasF32[4096]
-
blk.10.attn_q.weightF16[4096 4096]
-
blk.10.attn_v.biasF32[512]
-
blk.10.attn_v.weightF16[4096 512]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightF16[13440 4096]
-
blk.11.ffn_gate.weightF16[4096 13440]
-
blk.11.ffn_up.weightF16[4096 13440]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.biasF32[512]
-
blk.11.attn_k.weightF16[4096 512]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.attn_q.biasF32[4096]
-
blk.11.attn_q.weightF16[4096 4096]
-
blk.11.attn_v.biasF32[512]
-
blk.11.attn_v.weightF16[4096 512]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightF16[13440 4096]
-
blk.12.ffn_gate.weightF16[4096 13440]
-
blk.12.ffn_up.weightF16[4096 13440]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.biasF32[512]
-
blk.12.attn_k.weightF16[4096 512]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.attn_q.biasF32[4096]
-
blk.12.attn_q.weightF16[4096 4096]
-
blk.12.attn_v.biasF32[512]
-
blk.12.attn_v.weightF16[4096 512]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightF16[13440 4096]
-
blk.13.ffn_gate.weightF16[4096 13440]
-
blk.13.ffn_up.weightF16[4096 13440]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.biasF32[512]
-
blk.13.attn_k.weightF16[4096 512]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.attn_q.biasF32[4096]
-
blk.13.attn_q.weightF16[4096 4096]
-
blk.13.attn_v.biasF32[512]
-
blk.13.attn_v.weightF16[4096 512]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightF16[13440 4096]
-
blk.14.ffn_gate.weightF16[4096 13440]
-
blk.14.ffn_up.weightF16[4096 13440]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.biasF32[512]
-
blk.14.attn_k.weightF16[4096 512]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.attn_q.biasF32[4096]
-
blk.14.attn_q.weightF16[4096 4096]
-
blk.14.attn_v.biasF32[512]
-
blk.14.attn_v.weightF16[4096 512]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightF16[13440 4096]
-
blk.15.ffn_gate.weightF16[4096 13440]
-
blk.15.ffn_up.weightF16[4096 13440]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.biasF32[512]
-
blk.15.attn_k.weightF16[4096 512]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.attn_q.biasF32[4096]
-
blk.15.attn_q.weightF16[4096 4096]
-
blk.15.attn_v.biasF32[512]
-
blk.15.attn_v.weightF16[4096 512]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightF16[13440 4096]
-
blk.16.ffn_gate.weightF16[4096 13440]
-
blk.16.ffn_up.weightF16[4096 13440]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.biasF32[512]
-
blk.16.attn_k.weightF16[4096 512]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.attn_q.biasF32[4096]
-
blk.16.attn_q.weightF16[4096 4096]
-
blk.16.attn_v.biasF32[512]
-
blk.16.attn_v.weightF16[4096 512]
-
blk.17.ffn_gate.weightF16[4096 13440]
-
blk.17.attn_k.biasF32[512]
-
blk.17.attn_k.weightF16[4096 512]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.attn_q.biasF32[4096]
-
blk.17.attn_q.weightF16[4096 4096]
-
blk.17.attn_v.biasF32[512]
-
blk.17.attn_v.weightF16[4096 512]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightF16[13440 4096]
-
blk.17.ffn_up.weightF16[4096 13440]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightF16[13440 4096]
-
blk.18.ffn_gate.weightF16[4096 13440]
-
blk.18.ffn_up.weightF16[4096 13440]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.biasF32[512]
-
blk.18.attn_k.weightF16[4096 512]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.attn_q.biasF32[4096]
-
blk.18.attn_q.weightF16[4096 4096]
-
blk.18.attn_v.biasF32[512]
-
blk.18.attn_v.weightF16[4096 512]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightF16[13440 4096]
-
blk.19.ffn_gate.weightF16[4096 13440]
-
blk.19.ffn_up.weightF16[4096 13440]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.biasF32[512]
-
blk.19.attn_k.weightF16[4096 512]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.attn_q.biasF32[4096]
-
blk.19.attn_q.weightF16[4096 4096]
-
blk.19.attn_v.biasF32[512]
-
blk.19.attn_v.weightF16[4096 512]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightF16[13440 4096]
-
blk.20.ffn_gate.weightF16[4096 13440]
-
blk.20.ffn_up.weightF16[4096 13440]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.biasF32[512]
-
blk.20.attn_k.weightF16[4096 512]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.attn_q.biasF32[4096]
-
blk.20.attn_q.weightF16[4096 4096]
-
blk.20.attn_v.biasF32[512]
-
blk.20.attn_v.weightF16[4096 512]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightF16[13440 4096]
-
blk.21.ffn_gate.weightF16[4096 13440]
-
blk.21.ffn_up.weightF16[4096 13440]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.biasF32[512]
-
blk.21.attn_k.weightF16[4096 512]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.attn_q.biasF32[4096]
-
blk.21.attn_q.weightF16[4096 4096]
-
blk.21.attn_v.biasF32[512]
-
blk.21.attn_v.weightF16[4096 512]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightF16[13440 4096]
-
blk.22.ffn_gate.weightF16[4096 13440]
-
blk.22.ffn_up.weightF16[4096 13440]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.biasF32[512]
-
blk.22.attn_k.weightF16[4096 512]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.attn_q.biasF32[4096]
-
blk.22.attn_q.weightF16[4096 4096]
-
blk.22.attn_v.biasF32[512]
-
blk.22.attn_v.weightF16[4096 512]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightF16[13440 4096]
-
blk.23.ffn_gate.weightF16[4096 13440]
-
blk.23.ffn_up.weightF16[4096 13440]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.biasF32[512]
-
blk.23.attn_k.weightF16[4096 512]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.attn_q.biasF32[4096]
-
blk.23.attn_q.weightF16[4096 4096]
-
blk.23.attn_v.biasF32[512]
-
blk.23.attn_v.weightF16[4096 512]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightF16[13440 4096]
-
blk.24.ffn_gate.weightF16[4096 13440]
-
blk.24.ffn_up.weightF16[4096 13440]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.biasF32[512]
-
blk.24.attn_k.weightF16[4096 512]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.attn_q.biasF32[4096]
-
blk.24.attn_q.weightF16[4096 4096]
-
blk.24.attn_v.biasF32[512]
-
blk.24.attn_v.weightF16[4096 512]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightF16[13440 4096]
-
blk.25.ffn_gate.weightF16[4096 13440]
-
blk.25.ffn_up.weightF16[4096 13440]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.biasF32[512]
-
blk.25.attn_k.weightF16[4096 512]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.attn_q.biasF32[4096]
-
blk.25.attn_q.weightF16[4096 4096]
-
blk.25.attn_v.biasF32[512]
-
blk.25.attn_v.weightF16[4096 512]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightF16[13440 4096]
-
blk.26.ffn_gate.weightF16[4096 13440]
-
blk.26.ffn_up.weightF16[4096 13440]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.biasF32[512]
-
blk.26.attn_k.weightF16[4096 512]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.attn_q.biasF32[4096]
-
blk.26.attn_q.weightF16[4096 4096]
-
blk.26.attn_v.biasF32[512]
-
blk.26.attn_v.weightF16[4096 512]
-
blk.27.attn_k.biasF32[512]
-
blk.27.attn_k.weightF16[4096 512]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.27.attn_q.biasF32[4096]
-
blk.27.attn_q.weightF16[4096 4096]
-
blk.27.attn_v.biasF32[512]
-
blk.27.attn_v.weightF16[4096 512]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightF16[13440 4096]
-
blk.27.ffn_gate.weightF16[4096 13440]
-
blk.27.ffn_up.weightF16[4096 13440]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightF16[13440 4096]
-
blk.28.ffn_gate.weightF16[4096 13440]
-
blk.28.ffn_up.weightF16[4096 13440]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.biasF32[512]
-
blk.28.attn_k.weightF16[4096 512]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.attn_q.biasF32[4096]
-
blk.28.attn_q.weightF16[4096 4096]
-
blk.28.attn_v.biasF32[512]
-
blk.28.attn_v.weightF16[4096 512]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightF16[13440 4096]
-
blk.29.ffn_gate.weightF16[4096 13440]
-
blk.29.ffn_up.weightF16[4096 13440]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.biasF32[512]
-
blk.29.attn_k.weightF16[4096 512]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.attn_q.biasF32[4096]
-
blk.29.attn_q.weightF16[4096 4096]
-
blk.29.attn_v.biasF32[512]
-
blk.29.attn_v.weightF16[4096 512]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightF16[13440 4096]
-
blk.30.ffn_gate.weightF16[4096 13440]
-
blk.30.ffn_up.weightF16[4096 13440]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.biasF32[512]
-
blk.30.attn_k.weightF16[4096 512]
-
blk.30.attn_output.weightF16[4096 4096]
-
blk.30.attn_q.biasF32[4096]
-
blk.30.attn_q.weightF16[4096 4096]
-
blk.30.attn_v.biasF32[512]
-
blk.30.attn_v.weightF16[4096 512]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightF16[13440 4096]
-
blk.31.ffn_gate.weightF16[4096 13440]
-
blk.31.ffn_up.weightF16[4096 13440]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.biasF32[512]
-
blk.31.attn_k.weightF16[4096 512]
-
blk.31.attn_output.weightF16[4096 4096]
-
blk.31.attn_q.biasF32[4096]
-
blk.31.attn_q.weightF16[4096 4096]
-
blk.31.attn_v.biasF32[512]
-
blk.31.attn_v.weightF16[4096 512]
-
output.weightF16[4096 92416]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31