xe / yi-chat-6b
6B
22 Pulls Updated 9 months ago
2df38be6ce3c · 12GB
-
general.architecturellama
-
general.file_typeF16
-
llama.attention.head_count32
-
llama.attention.head_count_kv4
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length4096
-
llama.embedding_length4096
-
llama.feed_forward_length11008
-
llama.rope.dimension_count128
-
llama.rope.freq_base5e+06
-
tokenizer.ggml.add_bos_tokenfalse
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id2
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 3 3 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <|Human|> <|Assistant|> ...]
-
NameTypeShape
-
token_embd.weightF16[4096 64000]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightF16[11008 4096]
-
blk.0.ffn_gate.weightF16[4096 11008]
-
blk.0.ffn_up.weightF16[4096 11008]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightF16[4096 512]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.attn_q.weightF16[4096 4096]
-
blk.0.attn_v.weightF16[4096 512]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightF16[11008 4096]
-
blk.1.ffn_gate.weightF16[4096 11008]
-
blk.1.ffn_up.weightF16[4096 11008]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightF16[4096 512]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.attn_q.weightF16[4096 4096]
-
blk.1.attn_v.weightF16[4096 512]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightF16[11008 4096]
-
blk.2.ffn_gate.weightF16[4096 11008]
-
blk.2.ffn_up.weightF16[4096 11008]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightF16[4096 512]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.attn_q.weightF16[4096 4096]
-
blk.2.attn_v.weightF16[4096 512]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightF16[11008 4096]
-
blk.3.ffn_gate.weightF16[4096 11008]
-
blk.3.ffn_up.weightF16[4096 11008]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightF16[4096 512]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.attn_q.weightF16[4096 4096]
-
blk.3.attn_v.weightF16[4096 512]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightF16[11008 4096]
-
blk.4.ffn_gate.weightF16[4096 11008]
-
blk.4.ffn_up.weightF16[4096 11008]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightF16[4096 512]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.attn_q.weightF16[4096 4096]
-
blk.4.attn_v.weightF16[4096 512]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightF16[11008 4096]
-
blk.5.ffn_gate.weightF16[4096 11008]
-
blk.5.ffn_up.weightF16[4096 11008]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightF16[4096 512]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.5.attn_q.weightF16[4096 4096]
-
blk.5.attn_v.weightF16[4096 512]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightF16[11008 4096]
-
blk.6.ffn_gate.weightF16[4096 11008]
-
blk.6.ffn_up.weightF16[4096 11008]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightF16[4096 512]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.attn_q.weightF16[4096 4096]
-
blk.6.attn_v.weightF16[4096 512]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightF16[11008 4096]
-
blk.7.ffn_gate.weightF16[4096 11008]
-
blk.7.ffn_up.weightF16[4096 11008]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightF16[4096 512]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.attn_q.weightF16[4096 4096]
-
blk.7.attn_v.weightF16[4096 512]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightF16[11008 4096]
-
blk.8.ffn_gate.weightF16[4096 11008]
-
blk.8.ffn_up.weightF16[4096 11008]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightF16[4096 512]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.attn_q.weightF16[4096 4096]
-
blk.8.attn_v.weightF16[4096 512]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightF16[11008 4096]
-
blk.9.ffn_gate.weightF16[4096 11008]
-
blk.9.ffn_up.weightF16[4096 11008]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightF16[4096 512]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.attn_q.weightF16[4096 4096]
-
blk.9.attn_v.weightF16[4096 512]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightF16[11008 4096]
-
blk.10.ffn_gate.weightF16[4096 11008]
-
blk.10.ffn_up.weightF16[4096 11008]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.weightF16[4096 512]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.10.attn_q.weightF16[4096 4096]
-
blk.10.attn_v.weightF16[4096 512]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightF16[11008 4096]
-
blk.11.ffn_gate.weightF16[4096 11008]
-
blk.11.ffn_up.weightF16[4096 11008]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightF16[4096 512]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.attn_q.weightF16[4096 4096]
-
blk.11.attn_v.weightF16[4096 512]
-
blk.12.ffn_gate.weightF16[4096 11008]
-
blk.12.ffn_up.weightF16[4096 11008]
-
blk.12.attn_k.weightF16[4096 512]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.attn_q.weightF16[4096 4096]
-
blk.12.attn_v.weightF16[4096 512]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightF16[11008 4096]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightF16[11008 4096]
-
blk.13.ffn_gate.weightF16[4096 11008]
-
blk.13.ffn_up.weightF16[4096 11008]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightF16[4096 512]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.attn_q.weightF16[4096 4096]
-
blk.13.attn_v.weightF16[4096 512]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightF16[11008 4096]
-
blk.14.ffn_gate.weightF16[4096 11008]
-
blk.14.ffn_up.weightF16[4096 11008]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightF16[4096 512]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.attn_q.weightF16[4096 4096]
-
blk.14.attn_v.weightF16[4096 512]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightF16[11008 4096]
-
blk.15.ffn_gate.weightF16[4096 11008]
-
blk.15.ffn_up.weightF16[4096 11008]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightF16[4096 512]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.attn_q.weightF16[4096 4096]
-
blk.15.attn_v.weightF16[4096 512]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightF16[11008 4096]
-
blk.16.ffn_gate.weightF16[4096 11008]
-
blk.16.ffn_up.weightF16[4096 11008]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightF16[4096 512]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.attn_q.weightF16[4096 4096]
-
blk.16.attn_v.weightF16[4096 512]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightF16[11008 4096]
-
blk.17.ffn_gate.weightF16[4096 11008]
-
blk.17.ffn_up.weightF16[4096 11008]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightF16[4096 512]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.attn_q.weightF16[4096 4096]
-
blk.17.attn_v.weightF16[4096 512]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightF16[11008 4096]
-
blk.18.ffn_gate.weightF16[4096 11008]
-
blk.18.ffn_up.weightF16[4096 11008]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightF16[4096 512]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.attn_q.weightF16[4096 4096]
-
blk.18.attn_v.weightF16[4096 512]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightF16[11008 4096]
-
blk.19.ffn_gate.weightF16[4096 11008]
-
blk.19.ffn_up.weightF16[4096 11008]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightF16[4096 512]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.attn_q.weightF16[4096 4096]
-
blk.19.attn_v.weightF16[4096 512]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightF16[11008 4096]
-
blk.20.ffn_gate.weightF16[4096 11008]
-
blk.20.ffn_up.weightF16[4096 11008]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.weightF16[4096 512]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.attn_q.weightF16[4096 4096]
-
blk.20.attn_v.weightF16[4096 512]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightF16[11008 4096]
-
blk.21.ffn_gate.weightF16[4096 11008]
-
blk.21.ffn_up.weightF16[4096 11008]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightF16[4096 512]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.attn_q.weightF16[4096 4096]
-
blk.21.attn_v.weightF16[4096 512]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightF16[11008 4096]
-
blk.22.ffn_gate.weightF16[4096 11008]
-
blk.22.ffn_up.weightF16[4096 11008]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.weightF16[4096 512]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.attn_q.weightF16[4096 4096]
-
blk.22.attn_v.weightF16[4096 512]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightF16[11008 4096]
-
blk.23.ffn_gate.weightF16[4096 11008]
-
blk.23.ffn_up.weightF16[4096 11008]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightF16[4096 512]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.attn_q.weightF16[4096 4096]
-
blk.23.attn_v.weightF16[4096 512]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightF16[11008 4096]
-
blk.24.ffn_gate.weightF16[4096 11008]
-
blk.24.ffn_up.weightF16[4096 11008]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightF16[4096 512]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.attn_q.weightF16[4096 4096]
-
blk.24.attn_v.weightF16[4096 512]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightF16[11008 4096]
-
blk.25.ffn_gate.weightF16[4096 11008]
-
blk.25.ffn_up.weightF16[4096 11008]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightF16[4096 512]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.attn_q.weightF16[4096 4096]
-
blk.25.attn_v.weightF16[4096 512]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightF16[11008 4096]
-
blk.26.ffn_gate.weightF16[4096 11008]
-
blk.26.ffn_up.weightF16[4096 11008]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightF16[4096 512]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.attn_q.weightF16[4096 4096]
-
blk.26.attn_v.weightF16[4096 512]
-
blk.27.attn_k.weightF16[4096 512]
-
blk.27.attn_q.weightF16[4096 4096]
-
blk.27.attn_v.weightF16[4096 512]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightF16[11008 4096]
-
blk.27.ffn_gate.weightF16[4096 11008]
-
blk.27.ffn_up.weightF16[4096 11008]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightF16[11008 4096]
-
blk.28.ffn_gate.weightF16[4096 11008]
-
blk.28.ffn_up.weightF16[4096 11008]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightF16[4096 512]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.attn_q.weightF16[4096 4096]
-
blk.28.attn_v.weightF16[4096 512]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightF16[11008 4096]
-
blk.29.ffn_gate.weightF16[4096 11008]
-
blk.29.ffn_up.weightF16[4096 11008]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightF16[4096 512]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.attn_q.weightF16[4096 4096]
-
blk.29.attn_v.weightF16[4096 512]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightF16[11008 4096]
-
blk.30.ffn_gate.weightF16[4096 11008]
-
blk.30.ffn_up.weightF16[4096 11008]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightF16[4096 512]
-
blk.30.attn_output.weightF16[4096 4096]
-
blk.30.attn_q.weightF16[4096 4096]
-
blk.30.attn_v.weightF16[4096 512]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightF16[11008 4096]
-
blk.31.ffn_gate.weightF16[4096 11008]
-
blk.31.ffn_up.weightF16[4096 11008]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.weightF16[4096 512]
-
blk.31.attn_output.weightF16[4096 4096]
-
blk.31.attn_q.weightF16[4096 4096]
-
blk.31.attn_v.weightF16[4096 512]
-
output.weightF16[4096 64000]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31