latest
15GB
18 Pulls Updated 4 months ago
e09e0c468a34 · 15GB
-
general.architectureqwen
-
qwen.attention.head_count32
-
qwen.attention.layer_norm_rms_epsilon1e-06
-
qwen.block_count32
-
qwen.context_length8192
-
qwen.embedding_length4096
-
qwen.feed_forward_length22016
-
qwen.rope.dimension_count128
-
qwen.rope.freq_base10000
-
tokenizer.ggml.bos_token_id151643
-
tokenizer.ggml.eos_token_id151643
-
tokenizer.ggml.merges[Ġ Ġ ĠĠ ĠĠ i n Ġ t ĠĠĠĠ ĠĠĠĠ ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.token_type[1 1 1 1 1 ...]
-
tokenizer.ggml.tokens[! " # $ % ...]
-
tokenizer.ggml.unknown_token_id151643
-
NameTypeShape
-
token_embd.weightF16[4096 151936]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.attn_qkv.weightF16[4096 12288]
-
blk.0.attn_qkv.biasF32[12288]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.ffn_up.weightF16[4096 11008]
-
blk.0.ffn_gate.weightF16[4096 11008]
-
blk.0.ffn_down.weightF16[11008 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.attn_qkv.weightF16[4096 12288]
-
blk.1.attn_qkv.biasF32[12288]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.ffn_up.weightF16[4096 11008]
-
blk.1.ffn_gate.weightF16[4096 11008]
-
blk.1.ffn_down.weightF16[11008 4096]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.attn_qkv.weightF16[4096 12288]
-
blk.2.attn_qkv.biasF32[12288]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.ffn_up.weightF16[4096 11008]
-
blk.2.ffn_gate.weightF16[4096 11008]
-
blk.2.ffn_down.weightF16[11008 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.attn_qkv.weightF16[4096 12288]
-
blk.3.attn_qkv.biasF32[12288]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.ffn_up.weightF16[4096 11008]
-
blk.3.ffn_gate.weightF16[4096 11008]
-
blk.3.ffn_down.weightF16[11008 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.attn_qkv.weightF16[4096 12288]
-
blk.4.attn_qkv.biasF32[12288]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.ffn_up.weightF16[4096 11008]
-
blk.4.ffn_gate.weightF16[4096 11008]
-
blk.4.ffn_down.weightF16[11008 4096]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.attn_qkv.weightF16[4096 12288]
-
blk.5.attn_qkv.biasF32[12288]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.ffn_up.weightF16[4096 11008]
-
blk.5.ffn_gate.weightF16[4096 11008]
-
blk.5.ffn_down.weightF16[11008 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.attn_qkv.weightF16[4096 12288]
-
blk.6.attn_qkv.biasF32[12288]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.ffn_up.weightF16[4096 11008]
-
blk.6.ffn_gate.weightF16[4096 11008]
-
blk.6.ffn_down.weightF16[11008 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.attn_qkv.weightF16[4096 12288]
-
blk.7.attn_qkv.biasF32[12288]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.ffn_up.weightF16[4096 11008]
-
blk.7.ffn_gate.weightF16[4096 11008]
-
blk.7.ffn_down.weightF16[11008 4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.attn_qkv.weightF16[4096 12288]
-
blk.8.attn_qkv.biasF32[12288]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.ffn_up.weightF16[4096 11008]
-
blk.8.ffn_gate.weightF16[4096 11008]
-
blk.8.ffn_down.weightF16[11008 4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.attn_qkv.weightF16[4096 12288]
-
blk.9.attn_qkv.biasF32[12288]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.ffn_up.weightF16[4096 11008]
-
blk.9.ffn_gate.weightF16[4096 11008]
-
blk.9.ffn_down.weightF16[11008 4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.attn_qkv.weightF16[4096 12288]
-
blk.10.attn_qkv.biasF32[12288]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.ffn_up.weightF16[4096 11008]
-
blk.10.ffn_gate.weightF16[4096 11008]
-
blk.10.ffn_down.weightF16[11008 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.attn_qkv.weightF16[4096 12288]
-
blk.11.attn_qkv.biasF32[12288]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.ffn_up.weightF16[4096 11008]
-
blk.11.ffn_gate.weightF16[4096 11008]
-
blk.11.ffn_down.weightF16[11008 4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.attn_qkv.weightF16[4096 12288]
-
blk.12.attn_qkv.biasF32[12288]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.ffn_up.weightF16[4096 11008]
-
blk.12.ffn_gate.weightF16[4096 11008]
-
blk.12.ffn_down.weightF16[11008 4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.attn_qkv.weightF16[4096 12288]
-
blk.13.attn_qkv.biasF32[12288]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.ffn_up.weightF16[4096 11008]
-
blk.13.ffn_gate.weightF16[4096 11008]
-
blk.13.ffn_down.weightF16[11008 4096]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.attn_qkv.weightF16[4096 12288]
-
blk.14.attn_qkv.biasF32[12288]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.ffn_up.weightF16[4096 11008]
-
blk.14.ffn_gate.weightF16[4096 11008]
-
blk.14.ffn_down.weightF16[11008 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.attn_qkv.weightF16[4096 12288]
-
blk.15.attn_qkv.biasF32[12288]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.ffn_up.weightF16[4096 11008]
-
blk.15.ffn_gate.weightF16[4096 11008]
-
blk.15.ffn_down.weightF16[11008 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.attn_qkv.weightF16[4096 12288]
-
blk.16.attn_qkv.biasF32[12288]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.ffn_up.weightF16[4096 11008]
-
blk.16.ffn_gate.weightF16[4096 11008]
-
blk.16.ffn_down.weightF16[11008 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.attn_qkv.weightF16[4096 12288]
-
blk.17.attn_qkv.biasF32[12288]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.ffn_up.weightF16[4096 11008]
-
blk.17.ffn_gate.weightF16[4096 11008]
-
blk.17.ffn_down.weightF16[11008 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.attn_qkv.weightF16[4096 12288]
-
blk.18.attn_qkv.biasF32[12288]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.ffn_up.weightF16[4096 11008]
-
blk.18.ffn_gate.weightF16[4096 11008]
-
blk.18.ffn_down.weightF16[11008 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.attn_qkv.weightF16[4096 12288]
-
blk.19.attn_qkv.biasF32[12288]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.ffn_up.weightF16[4096 11008]
-
blk.19.ffn_gate.weightF16[4096 11008]
-
blk.19.ffn_down.weightF16[11008 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.attn_qkv.weightF16[4096 12288]
-
blk.20.attn_qkv.biasF32[12288]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.ffn_up.weightF16[4096 11008]
-
blk.20.ffn_gate.weightF16[4096 11008]
-
blk.20.ffn_down.weightF16[11008 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.attn_qkv.weightF16[4096 12288]
-
blk.21.attn_qkv.biasF32[12288]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.ffn_up.weightF16[4096 11008]
-
blk.21.ffn_gate.weightF16[4096 11008]
-
blk.21.ffn_down.weightF16[11008 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.attn_qkv.weightF16[4096 12288]
-
blk.22.attn_qkv.biasF32[12288]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.ffn_up.weightF16[4096 11008]
-
blk.22.ffn_gate.weightF16[4096 11008]
-
blk.22.ffn_down.weightF16[11008 4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.attn_qkv.weightF16[4096 12288]
-
blk.23.attn_qkv.biasF32[12288]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.ffn_up.weightF16[4096 11008]
-
blk.23.ffn_gate.weightF16[4096 11008]
-
blk.23.ffn_down.weightF16[11008 4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.attn_qkv.weightF16[4096 12288]
-
blk.24.attn_qkv.biasF32[12288]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.ffn_up.weightF16[4096 11008]
-
blk.24.ffn_gate.weightF16[4096 11008]
-
blk.24.ffn_down.weightF16[11008 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.attn_qkv.weightF16[4096 12288]
-
blk.25.attn_qkv.biasF32[12288]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.ffn_up.weightF16[4096 11008]
-
blk.25.ffn_gate.weightF16[4096 11008]
-
blk.25.ffn_down.weightF16[11008 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.attn_qkv.weightF16[4096 12288]
-
blk.26.attn_qkv.biasF32[12288]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.ffn_up.weightF16[4096 11008]
-
blk.26.ffn_gate.weightF16[4096 11008]
-
blk.26.ffn_down.weightF16[11008 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.attn_qkv.weightF16[4096 12288]
-
blk.27.attn_qkv.biasF32[12288]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.ffn_up.weightF16[4096 11008]
-
blk.27.ffn_gate.weightF16[4096 11008]
-
blk.27.ffn_down.weightF16[11008 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.attn_qkv.weightF16[4096 12288]
-
blk.28.attn_qkv.biasF32[12288]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.ffn_up.weightF16[4096 11008]
-
blk.28.ffn_gate.weightF16[4096 11008]
-
blk.28.ffn_down.weightF16[11008 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.attn_qkv.weightF16[4096 12288]
-
blk.29.attn_qkv.biasF32[12288]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.ffn_up.weightF16[4096 11008]
-
blk.29.ffn_gate.weightF16[4096 11008]
-
blk.29.ffn_down.weightF16[11008 4096]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.attn_qkv.weightF16[4096 12288]
-
blk.30.attn_qkv.biasF32[12288]
-
blk.30.attn_output.weightF16[4096 4096]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.ffn_up.weightF16[4096 11008]
-
blk.30.ffn_gate.weightF16[4096 11008]
-
blk.30.ffn_down.weightF16[11008 4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.attn_qkv.weightF16[4096 12288]
-
blk.31.attn_qkv.biasF32[12288]
-
blk.31.attn_output.weightF16[4096 4096]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.ffn_up.weightF16[4096 11008]
-
blk.31.ffn_gate.weightF16[4096 11008]
-
blk.31.ffn_down.weightF16[11008 4096]
-
output_norm.weightF32[4096]
-
output.weightF16[4096 151936]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31