latest
15GB
7B
1 Pull Updated 2 months ago
490adf80d204 · 15GB
-
general.architectureqwen2
-
general.file_typeF16
-
qwen2.attention.head_count32
-
qwen2.attention.head_count_kv4
-
qwen2.attention.layer_norm_rms_epsilon1e-05
-
qwen2.block_count32
-
qwen2.context_length65536
-
qwen2.embedding_length4096
-
qwen2.feed_forward_length13440
-
qwen2.rope.freq_base1e+06
-
tokenizer.ggml.add_bos_tokenfalse
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id2
-
tokenizer.ggml.eos_token_id4
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.padding_token_id92298
-
tokenizer.ggml.predefault
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 3 3 ...]
-
tokenizer.ggml.tokens[<unk> <s> <|endoftext|> <|im_start|> <|im_end|> ...]
-
tokenizer.ggml.unknown_token_id0
-
NameTypeShape
-
token_embd.weightF16[4096 92416]
-
blk.0.ffn_gate.weightF16[4096 13440]
-
blk.0.attn_k.biasF32[512]
-
blk.0.attn_k.weightF16[4096 512]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.attn_q.biasF32[4096]
-
blk.0.attn_q.weightF16[4096 4096]
-
blk.0.attn_v.biasF32[512]
-
blk.0.attn_v.weightF16[4096 512]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightF16[13440 4096]
-
blk.0.ffn_up.weightF16[4096 13440]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightF16[13440 4096]
-
blk.1.ffn_gate.weightF16[4096 13440]
-
blk.1.ffn_up.weightF16[4096 13440]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.biasF32[512]
-
blk.1.attn_k.weightF16[4096 512]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.attn_q.biasF32[4096]
-
blk.1.attn_q.weightF16[4096 4096]
-
blk.1.attn_v.biasF32[512]
-
blk.1.attn_v.weightF16[4096 512]
-
blk.2.ffn_gate.weightF16[4096 13440]
-
blk.2.ffn_up.weightF16[4096 13440]
-
blk.2.attn_k.biasF32[512]
-
blk.2.attn_k.weightF16[4096 512]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.attn_q.biasF32[4096]
-
blk.2.attn_q.weightF16[4096 4096]
-
blk.2.attn_v.biasF32[512]
-
blk.2.attn_v.weightF16[4096 512]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightF16[13440 4096]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightF16[13440 4096]
-
blk.3.ffn_gate.weightF16[4096 13440]
-
blk.3.ffn_up.weightF16[4096 13440]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.biasF32[512]
-
blk.3.attn_k.weightF16[4096 512]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.attn_q.biasF32[4096]
-
blk.3.attn_q.weightF16[4096 4096]
-
blk.3.attn_v.biasF32[512]
-
blk.3.attn_v.weightF16[4096 512]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightF16[13440 4096]
-
blk.4.ffn_gate.weightF16[4096 13440]
-
blk.4.ffn_up.weightF16[4096 13440]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.biasF32[512]
-
blk.4.attn_k.weightF16[4096 512]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.attn_q.biasF32[4096]
-
blk.4.attn_q.weightF16[4096 4096]
-
blk.4.attn_v.biasF32[512]
-
blk.4.attn_v.weightF16[4096 512]
-
blk.5.attn_k.biasF32[512]
-
blk.5.attn_k.weightF16[4096 512]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.5.attn_q.biasF32[4096]
-
blk.5.attn_q.weightF16[4096 4096]
-
blk.5.attn_v.biasF32[512]
-
blk.5.attn_v.weightF16[4096 512]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightF16[13440 4096]
-
blk.5.ffn_gate.weightF16[4096 13440]
-
blk.5.ffn_up.weightF16[4096 13440]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightF16[13440 4096]
-
blk.6.ffn_gate.weightF16[4096 13440]
-
blk.6.ffn_up.weightF16[4096 13440]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.biasF32[512]
-
blk.6.attn_k.weightF16[4096 512]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.attn_q.biasF32[4096]
-
blk.6.attn_q.weightF16[4096 4096]
-
blk.6.attn_v.biasF32[512]
-
blk.6.attn_v.weightF16[4096 512]
-
blk.7.ffn_gate.weightF16[4096 13440]
-
blk.7.attn_k.biasF32[512]
-
blk.7.attn_k.weightF16[4096 512]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.attn_q.biasF32[4096]
-
blk.7.attn_q.weightF16[4096 4096]
-
blk.7.attn_v.biasF32[512]
-
blk.7.attn_v.weightF16[4096 512]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightF16[13440 4096]
-
blk.7.ffn_up.weightF16[4096 13440]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightF16[13440 4096]
-
blk.8.ffn_gate.weightF16[4096 13440]
-
blk.8.ffn_up.weightF16[4096 13440]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.biasF32[512]
-
blk.8.attn_k.weightF16[4096 512]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.attn_q.biasF32[4096]
-
blk.8.attn_q.weightF16[4096 4096]
-
blk.8.attn_v.biasF32[512]
-
blk.8.attn_v.weightF16[4096 512]
-
blk.9.ffn_gate.weightF16[4096 13440]
-
blk.9.ffn_up.weightF16[4096 13440]
-
blk.9.attn_k.biasF32[512]
-
blk.9.attn_k.weightF16[4096 512]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.attn_q.biasF32[4096]
-
blk.9.attn_q.weightF16[4096 4096]
-
blk.9.attn_v.biasF32[512]
-
blk.9.attn_v.weightF16[4096 512]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightF16[13440 4096]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightF16[13440 4096]
-
blk.10.ffn_gate.weightF16[4096 13440]
-
blk.10.ffn_up.weightF16[4096 13440]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.biasF32[512]
-
blk.10.attn_k.weightF16[4096 512]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.10.attn_q.biasF32[4096]
-
blk.10.attn_q.weightF16[4096 4096]
-
blk.10.attn_v.biasF32[512]
-
blk.10.attn_v.weightF16[4096 512]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightF16[13440 4096]
-
blk.11.ffn_gate.weightF16[4096 13440]
-
blk.11.ffn_up.weightF16[4096 13440]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.biasF32[512]
-
blk.11.attn_k.weightF16[4096 512]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.attn_q.biasF32[4096]
-
blk.11.attn_q.weightF16[4096 4096]
-
blk.11.attn_v.biasF32[512]
-
blk.11.attn_v.weightF16[4096 512]
-
blk.12.attn_k.biasF32[512]
-
blk.12.attn_k.weightF16[4096 512]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.attn_q.biasF32[4096]
-
blk.12.attn_q.weightF16[4096 4096]
-
blk.12.attn_v.biasF32[512]
-
blk.12.attn_v.weightF16[4096 512]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightF16[13440 4096]
-
blk.12.ffn_gate.weightF16[4096 13440]
-
blk.12.ffn_up.weightF16[4096 13440]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightF16[13440 4096]
-
blk.13.ffn_gate.weightF16[4096 13440]
-
blk.13.ffn_up.weightF16[4096 13440]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.biasF32[512]
-
blk.13.attn_k.weightF16[4096 512]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.attn_q.biasF32[4096]
-
blk.13.attn_q.weightF16[4096 4096]
-
blk.13.attn_v.biasF32[512]
-
blk.13.attn_v.weightF16[4096 512]
-
blk.14.ffn_gate.weightF16[4096 13440]
-
blk.14.attn_k.biasF32[512]
-
blk.14.attn_k.weightF16[4096 512]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.attn_q.biasF32[4096]
-
blk.14.attn_q.weightF16[4096 4096]
-
blk.14.attn_v.biasF32[512]
-
blk.14.attn_v.weightF16[4096 512]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightF16[13440 4096]
-
blk.14.ffn_up.weightF16[4096 13440]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightF16[13440 4096]
-
blk.15.ffn_gate.weightF16[4096 13440]
-
blk.15.ffn_up.weightF16[4096 13440]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.biasF32[512]
-
blk.15.attn_k.weightF16[4096 512]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.attn_q.biasF32[4096]
-
blk.15.attn_q.weightF16[4096 4096]
-
blk.15.attn_v.biasF32[512]
-
blk.15.attn_v.weightF16[4096 512]
-
blk.16.ffn_gate.weightF16[4096 13440]
-
blk.16.ffn_up.weightF16[4096 13440]
-
blk.16.attn_k.biasF32[512]
-
blk.16.attn_k.weightF16[4096 512]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.attn_q.biasF32[4096]
-
blk.16.attn_q.weightF16[4096 4096]
-
blk.16.attn_v.biasF32[512]
-
blk.16.attn_v.weightF16[4096 512]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightF16[13440 4096]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightF16[13440 4096]
-
blk.17.ffn_gate.weightF16[4096 13440]
-
blk.17.ffn_up.weightF16[4096 13440]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.biasF32[512]
-
blk.17.attn_k.weightF16[4096 512]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.attn_q.biasF32[4096]
-
blk.17.attn_q.weightF16[4096 4096]
-
blk.17.attn_v.biasF32[512]
-
blk.17.attn_v.weightF16[4096 512]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightF16[13440 4096]
-
blk.18.ffn_gate.weightF16[4096 13440]
-
blk.18.ffn_up.weightF16[4096 13440]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.biasF32[512]
-
blk.18.attn_k.weightF16[4096 512]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.attn_q.biasF32[4096]
-
blk.18.attn_q.weightF16[4096 4096]
-
blk.18.attn_v.biasF32[512]
-
blk.18.attn_v.weightF16[4096 512]
-
blk.19.attn_k.biasF32[512]
-
blk.19.attn_k.weightF16[4096 512]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.attn_q.biasF32[4096]
-
blk.19.attn_q.weightF16[4096 4096]
-
blk.19.attn_v.biasF32[512]
-
blk.19.attn_v.weightF16[4096 512]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightF16[13440 4096]
-
blk.19.ffn_gate.weightF16[4096 13440]
-
blk.19.ffn_up.weightF16[4096 13440]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightF16[13440 4096]
-
blk.20.ffn_gate.weightF16[4096 13440]
-
blk.20.ffn_up.weightF16[4096 13440]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.biasF32[512]
-
blk.20.attn_k.weightF16[4096 512]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.attn_q.biasF32[4096]
-
blk.20.attn_q.weightF16[4096 4096]
-
blk.20.attn_v.biasF32[512]
-
blk.20.attn_v.weightF16[4096 512]
-
blk.21.ffn_gate.weightF16[4096 13440]
-
blk.21.attn_k.biasF32[512]
-
blk.21.attn_k.weightF16[4096 512]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.attn_q.biasF32[4096]
-
blk.21.attn_q.weightF16[4096 4096]
-
blk.21.attn_v.biasF32[512]
-
blk.21.attn_v.weightF16[4096 512]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightF16[13440 4096]
-
blk.21.ffn_up.weightF16[4096 13440]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightF16[13440 4096]
-
blk.22.ffn_gate.weightF16[4096 13440]
-
blk.22.ffn_up.weightF16[4096 13440]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.biasF32[512]
-
blk.22.attn_k.weightF16[4096 512]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.attn_q.biasF32[4096]
-
blk.22.attn_q.weightF16[4096 4096]
-
blk.22.attn_v.biasF32[512]
-
blk.22.attn_v.weightF16[4096 512]
-
blk.23.ffn_gate.weightF16[4096 13440]
-
blk.23.ffn_up.weightF16[4096 13440]
-
blk.23.attn_k.biasF32[512]
-
blk.23.attn_k.weightF16[4096 512]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.attn_q.biasF32[4096]
-
blk.23.attn_q.weightF16[4096 4096]
-
blk.23.attn_v.biasF32[512]
-
blk.23.attn_v.weightF16[4096 512]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightF16[13440 4096]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightF16[13440 4096]
-
blk.24.ffn_gate.weightF16[4096 13440]
-
blk.24.ffn_up.weightF16[4096 13440]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.biasF32[512]
-
blk.24.attn_k.weightF16[4096 512]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.attn_q.biasF32[4096]
-
blk.24.attn_q.weightF16[4096 4096]
-
blk.24.attn_v.biasF32[512]
-
blk.24.attn_v.weightF16[4096 512]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightF16[13440 4096]
-
blk.25.ffn_gate.weightF16[4096 13440]
-
blk.25.ffn_up.weightF16[4096 13440]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.biasF32[512]
-
blk.25.attn_k.weightF16[4096 512]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.attn_q.biasF32[4096]
-
blk.25.attn_q.weightF16[4096 4096]
-
blk.25.attn_v.biasF32[512]
-
blk.25.attn_v.weightF16[4096 512]
-
blk.26.attn_k.biasF32[512]
-
blk.26.attn_k.weightF16[4096 512]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.attn_q.biasF32[4096]
-
blk.26.attn_q.weightF16[4096 4096]
-
blk.26.attn_v.biasF32[512]
-
blk.26.attn_v.weightF16[4096 512]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightF16[13440 4096]
-
blk.26.ffn_gate.weightF16[4096 13440]
-
blk.26.ffn_up.weightF16[4096 13440]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightF16[13440 4096]
-
blk.27.ffn_gate.weightF16[4096 13440]
-
blk.27.ffn_up.weightF16[4096 13440]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.biasF32[512]
-
blk.27.attn_k.weightF16[4096 512]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.27.attn_q.biasF32[4096]
-
blk.27.attn_q.weightF16[4096 4096]
-
blk.27.attn_v.biasF32[512]
-
blk.27.attn_v.weightF16[4096 512]
-
blk.28.ffn_gate.weightF16[4096 13440]
-
blk.28.attn_k.biasF32[512]
-
blk.28.attn_k.weightF16[4096 512]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.attn_q.biasF32[4096]
-
blk.28.attn_q.weightF16[4096 4096]
-
blk.28.attn_v.biasF32[512]
-
blk.28.attn_v.weightF16[4096 512]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightF16[13440 4096]
-
blk.28.ffn_up.weightF16[4096 13440]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightF16[13440 4096]
-
blk.29.ffn_gate.weightF16[4096 13440]
-
blk.29.ffn_up.weightF16[4096 13440]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.biasF32[512]
-
blk.29.attn_k.weightF16[4096 512]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.attn_q.biasF32[4096]
-
blk.29.attn_q.weightF16[4096 4096]
-
blk.29.attn_v.biasF32[512]
-
blk.29.attn_v.weightF16[4096 512]
-
blk.30.ffn_gate.weightF16[4096 13440]
-
blk.30.ffn_up.weightF16[4096 13440]
-
blk.30.attn_k.biasF32[512]
-
blk.30.attn_k.weightF16[4096 512]
-
blk.30.attn_output.weightF16[4096 4096]
-
blk.30.attn_q.biasF32[4096]
-
blk.30.attn_q.weightF16[4096 4096]
-
blk.30.attn_v.biasF32[512]
-
blk.30.attn_v.weightF16[4096 512]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightF16[13440 4096]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightF16[13440 4096]
-
blk.31.ffn_gate.weightF16[4096 13440]
-
blk.31.ffn_up.weightF16[4096 13440]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.biasF32[512]
-
blk.31.attn_k.weightF16[4096 512]
-
blk.31.attn_output.weightF16[4096 4096]
-
blk.31.attn_q.biasF32[4096]
-
blk.31.attn_q.weightF16[4096 4096]
-
blk.31.attn_v.biasF32[512]
-
blk.31.attn_v.weightF16[4096 512]
-
output_norm.weightF32[4096]
-
output.weightF16[4096 92416]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31