latest
726MB
smaller llms fine tuned to be a voice assistant
30 Pulls Updated 5 weeks ago
c131de136f10 · 726MB
-
general.architecturellama
-
general.file_typeF16
-
llama.attention.head_count15
-
llama.attention.head_count_kv5
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length2048
-
llama.embedding_length960
-
llama.feed_forward_length2560
-
llama.rope.dimension_count64
-
llama.rope.freq_base10000
-
llama.vocab_size49152
-
tokenizer.ggml.add_bos_tokenfalse
-
tokenizer.ggml.add_space_prefixfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id2
-
tokenizer.ggml.merges[Ġ t Ġ a i n h e Ġ Ġ ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.padding_token_id0
-
tokenizer.ggml.presmollm
-
tokenizer.ggml.token_type[3 3 3 3 3 ...]
-
tokenizer.ggml.tokens[<|endoftext|> <|im_start|> <|im_end|> <repo_name> <reponame> ...]
-
tokenizer.ggml.unknown_token_id0
-
NameTypeShape
-
token_embd.weightF16[960 49152]
-
blk.0.attn_norm.weightF32[960]
-
blk.0.ffn_down.weightF16[2560 960]
-
blk.0.ffn_gate.weightF16[960 2560]
-
blk.0.ffn_up.weightF16[960 2560]
-
blk.0.ffn_norm.weightF32[960]
-
blk.0.attn_k.weightF16[960 320]
-
blk.0.attn_output.weightF16[960 960]
-
blk.0.attn_q.weightF16[960 960]
-
blk.0.attn_v.weightF16[960 320]
-
blk.1.attn_norm.weightF32[960]
-
blk.1.ffn_down.weightF16[2560 960]
-
blk.1.ffn_gate.weightF16[960 2560]
-
blk.1.ffn_up.weightF16[960 2560]
-
blk.1.ffn_norm.weightF32[960]
-
blk.1.attn_k.weightF16[960 320]
-
blk.1.attn_output.weightF16[960 960]
-
blk.1.attn_q.weightF16[960 960]
-
blk.1.attn_v.weightF16[960 320]
-
blk.2.attn_norm.weightF32[960]
-
blk.2.ffn_down.weightF16[2560 960]
-
blk.2.ffn_gate.weightF16[960 2560]
-
blk.2.ffn_up.weightF16[960 2560]
-
blk.2.ffn_norm.weightF32[960]
-
blk.2.attn_k.weightF16[960 320]
-
blk.2.attn_output.weightF16[960 960]
-
blk.2.attn_q.weightF16[960 960]
-
blk.2.attn_v.weightF16[960 320]
-
blk.3.attn_norm.weightF32[960]
-
blk.3.ffn_down.weightF16[2560 960]
-
blk.3.ffn_gate.weightF16[960 2560]
-
blk.3.ffn_up.weightF16[960 2560]
-
blk.3.ffn_norm.weightF32[960]
-
blk.3.attn_k.weightF16[960 320]
-
blk.3.attn_output.weightF16[960 960]
-
blk.3.attn_q.weightF16[960 960]
-
blk.3.attn_v.weightF16[960 320]
-
blk.4.attn_norm.weightF32[960]
-
blk.4.ffn_down.weightF16[2560 960]
-
blk.4.ffn_gate.weightF16[960 2560]
-
blk.4.ffn_up.weightF16[960 2560]
-
blk.4.ffn_norm.weightF32[960]
-
blk.4.attn_k.weightF16[960 320]
-
blk.4.attn_output.weightF16[960 960]
-
blk.4.attn_q.weightF16[960 960]
-
blk.4.attn_v.weightF16[960 320]
-
blk.5.attn_norm.weightF32[960]
-
blk.5.ffn_down.weightF16[2560 960]
-
blk.5.ffn_gate.weightF16[960 2560]
-
blk.5.ffn_up.weightF16[960 2560]
-
blk.5.ffn_norm.weightF32[960]
-
blk.5.attn_k.weightF16[960 320]
-
blk.5.attn_output.weightF16[960 960]
-
blk.5.attn_q.weightF16[960 960]
-
blk.5.attn_v.weightF16[960 320]
-
blk.6.attn_norm.weightF32[960]
-
blk.6.ffn_down.weightF16[2560 960]
-
blk.6.ffn_gate.weightF16[960 2560]
-
blk.6.ffn_up.weightF16[960 2560]
-
blk.6.ffn_norm.weightF32[960]
-
blk.6.attn_k.weightF16[960 320]
-
blk.6.attn_output.weightF16[960 960]
-
blk.6.attn_q.weightF16[960 960]
-
blk.6.attn_v.weightF16[960 320]
-
blk.7.attn_norm.weightF32[960]
-
blk.7.ffn_down.weightF16[2560 960]
-
blk.7.ffn_gate.weightF16[960 2560]
-
blk.7.ffn_up.weightF16[960 2560]
-
blk.7.ffn_norm.weightF32[960]
-
blk.7.attn_k.weightF16[960 320]
-
blk.7.attn_output.weightF16[960 960]
-
blk.7.attn_q.weightF16[960 960]
-
blk.7.attn_v.weightF16[960 320]
-
blk.8.attn_norm.weightF32[960]
-
blk.8.ffn_down.weightF16[2560 960]
-
blk.8.ffn_gate.weightF16[960 2560]
-
blk.8.ffn_up.weightF16[960 2560]
-
blk.8.ffn_norm.weightF32[960]
-
blk.8.attn_k.weightF16[960 320]
-
blk.8.attn_output.weightF16[960 960]
-
blk.8.attn_q.weightF16[960 960]
-
blk.8.attn_v.weightF16[960 320]
-
blk.9.attn_norm.weightF32[960]
-
blk.9.ffn_down.weightF16[2560 960]
-
blk.9.ffn_gate.weightF16[960 2560]
-
blk.9.ffn_up.weightF16[960 2560]
-
blk.9.ffn_norm.weightF32[960]
-
blk.9.attn_k.weightF16[960 320]
-
blk.9.attn_output.weightF16[960 960]
-
blk.9.attn_q.weightF16[960 960]
-
blk.9.attn_v.weightF16[960 320]
-
blk.10.attn_norm.weightF32[960]
-
blk.10.ffn_down.weightF16[2560 960]
-
blk.10.ffn_gate.weightF16[960 2560]
-
blk.10.ffn_up.weightF16[960 2560]
-
blk.10.ffn_norm.weightF32[960]
-
blk.10.attn_k.weightF16[960 320]
-
blk.10.attn_output.weightF16[960 960]
-
blk.10.attn_q.weightF16[960 960]
-
blk.10.attn_v.weightF16[960 320]
-
blk.11.attn_norm.weightF32[960]
-
blk.11.ffn_down.weightF16[2560 960]
-
blk.11.ffn_gate.weightF16[960 2560]
-
blk.11.ffn_up.weightF16[960 2560]
-
blk.11.ffn_norm.weightF32[960]
-
blk.11.attn_k.weightF16[960 320]
-
blk.11.attn_output.weightF16[960 960]
-
blk.11.attn_q.weightF16[960 960]
-
blk.11.attn_v.weightF16[960 320]
-
blk.12.attn_norm.weightF32[960]
-
blk.12.ffn_down.weightF16[2560 960]
-
blk.12.ffn_gate.weightF16[960 2560]
-
blk.12.ffn_up.weightF16[960 2560]
-
blk.12.ffn_norm.weightF32[960]
-
blk.12.attn_k.weightF16[960 320]
-
blk.12.attn_output.weightF16[960 960]
-
blk.12.attn_q.weightF16[960 960]
-
blk.12.attn_v.weightF16[960 320]
-
blk.13.attn_norm.weightF32[960]
-
blk.13.ffn_down.weightF16[2560 960]
-
blk.13.ffn_gate.weightF16[960 2560]
-
blk.13.ffn_up.weightF16[960 2560]
-
blk.13.ffn_norm.weightF32[960]
-
blk.13.attn_k.weightF16[960 320]
-
blk.13.attn_output.weightF16[960 960]
-
blk.13.attn_q.weightF16[960 960]
-
blk.13.attn_v.weightF16[960 320]
-
blk.14.attn_norm.weightF32[960]
-
blk.14.ffn_down.weightF16[2560 960]
-
blk.14.ffn_gate.weightF16[960 2560]
-
blk.14.ffn_up.weightF16[960 2560]
-
blk.14.ffn_norm.weightF32[960]
-
blk.14.attn_k.weightF16[960 320]
-
blk.14.attn_output.weightF16[960 960]
-
blk.14.attn_q.weightF16[960 960]
-
blk.14.attn_v.weightF16[960 320]
-
blk.15.attn_norm.weightF32[960]
-
blk.15.ffn_down.weightF16[2560 960]
-
blk.15.ffn_gate.weightF16[960 2560]
-
blk.15.ffn_up.weightF16[960 2560]
-
blk.15.ffn_norm.weightF32[960]
-
blk.15.attn_k.weightF16[960 320]
-
blk.15.attn_output.weightF16[960 960]
-
blk.15.attn_q.weightF16[960 960]
-
blk.15.attn_v.weightF16[960 320]
-
blk.16.attn_norm.weightF32[960]
-
blk.16.ffn_down.weightF16[2560 960]
-
blk.16.ffn_gate.weightF16[960 2560]
-
blk.16.ffn_up.weightF16[960 2560]
-
blk.16.ffn_norm.weightF32[960]
-
blk.16.attn_k.weightF16[960 320]
-
blk.16.attn_output.weightF16[960 960]
-
blk.16.attn_q.weightF16[960 960]
-
blk.16.attn_v.weightF16[960 320]
-
blk.17.attn_norm.weightF32[960]
-
blk.17.ffn_down.weightF16[2560 960]
-
blk.17.ffn_gate.weightF16[960 2560]
-
blk.17.ffn_up.weightF16[960 2560]
-
blk.17.ffn_norm.weightF32[960]
-
blk.17.attn_k.weightF16[960 320]
-
blk.17.attn_output.weightF16[960 960]
-
blk.17.attn_q.weightF16[960 960]
-
blk.17.attn_v.weightF16[960 320]
-
blk.18.attn_norm.weightF32[960]
-
blk.18.ffn_down.weightF16[2560 960]
-
blk.18.ffn_gate.weightF16[960 2560]
-
blk.18.ffn_up.weightF16[960 2560]
-
blk.18.ffn_norm.weightF32[960]
-
blk.18.attn_k.weightF16[960 320]
-
blk.18.attn_output.weightF16[960 960]
-
blk.18.attn_q.weightF16[960 960]
-
blk.18.attn_v.weightF16[960 320]
-
blk.19.attn_norm.weightF32[960]
-
blk.19.ffn_down.weightF16[2560 960]
-
blk.19.ffn_gate.weightF16[960 2560]
-
blk.19.ffn_up.weightF16[960 2560]
-
blk.19.ffn_norm.weightF32[960]
-
blk.19.attn_k.weightF16[960 320]
-
blk.19.attn_output.weightF16[960 960]
-
blk.19.attn_q.weightF16[960 960]
-
blk.19.attn_v.weightF16[960 320]
-
blk.20.attn_norm.weightF32[960]
-
blk.20.ffn_down.weightF16[2560 960]
-
blk.20.ffn_gate.weightF16[960 2560]
-
blk.20.ffn_up.weightF16[960 2560]
-
blk.20.ffn_norm.weightF32[960]
-
blk.20.attn_k.weightF16[960 320]
-
blk.20.attn_output.weightF16[960 960]
-
blk.20.attn_q.weightF16[960 960]
-
blk.20.attn_v.weightF16[960 320]
-
blk.21.attn_norm.weightF32[960]
-
blk.21.ffn_down.weightF16[2560 960]
-
blk.21.ffn_gate.weightF16[960 2560]
-
blk.21.ffn_up.weightF16[960 2560]
-
blk.21.ffn_norm.weightF32[960]
-
blk.21.attn_k.weightF16[960 320]
-
blk.21.attn_output.weightF16[960 960]
-
blk.21.attn_q.weightF16[960 960]
-
blk.21.attn_v.weightF16[960 320]
-
blk.22.attn_norm.weightF32[960]
-
blk.22.ffn_down.weightF16[2560 960]
-
blk.22.ffn_gate.weightF16[960 2560]
-
blk.22.ffn_up.weightF16[960 2560]
-
blk.22.ffn_norm.weightF32[960]
-
blk.22.attn_k.weightF16[960 320]
-
blk.22.attn_output.weightF16[960 960]
-
blk.22.attn_q.weightF16[960 960]
-
blk.22.attn_v.weightF16[960 320]
-
blk.23.attn_norm.weightF32[960]
-
blk.23.ffn_down.weightF16[2560 960]
-
blk.23.ffn_gate.weightF16[960 2560]
-
blk.23.ffn_up.weightF16[960 2560]
-
blk.23.ffn_norm.weightF32[960]
-
blk.23.attn_k.weightF16[960 320]
-
blk.23.attn_output.weightF16[960 960]
-
blk.23.attn_q.weightF16[960 960]
-
blk.23.attn_v.weightF16[960 320]
-
blk.24.attn_norm.weightF32[960]
-
blk.24.ffn_down.weightF16[2560 960]
-
blk.24.ffn_gate.weightF16[960 2560]
-
blk.24.ffn_up.weightF16[960 2560]
-
blk.24.ffn_norm.weightF32[960]
-
blk.24.attn_k.weightF16[960 320]
-
blk.24.attn_output.weightF16[960 960]
-
blk.24.attn_q.weightF16[960 960]
-
blk.24.attn_v.weightF16[960 320]
-
blk.25.attn_norm.weightF32[960]
-
blk.25.ffn_down.weightF16[2560 960]
-
blk.25.ffn_gate.weightF16[960 2560]
-
blk.25.ffn_up.weightF16[960 2560]
-
blk.25.ffn_norm.weightF32[960]
-
blk.25.attn_k.weightF16[960 320]
-
blk.25.attn_output.weightF16[960 960]
-
blk.25.attn_q.weightF16[960 960]
-
blk.25.attn_v.weightF16[960 320]
-
blk.26.attn_norm.weightF32[960]
-
blk.26.ffn_down.weightF16[2560 960]
-
blk.26.ffn_gate.weightF16[960 2560]
-
blk.26.ffn_up.weightF16[960 2560]
-
blk.26.ffn_norm.weightF32[960]
-
blk.26.attn_k.weightF16[960 320]
-
blk.26.attn_output.weightF16[960 960]
-
blk.26.attn_q.weightF16[960 960]
-
blk.26.attn_v.weightF16[960 320]
-
blk.27.attn_norm.weightF32[960]
-
blk.27.ffn_down.weightF16[2560 960]
-
blk.27.ffn_gate.weightF16[960 2560]
-
blk.27.ffn_up.weightF16[960 2560]
-
blk.27.ffn_norm.weightF32[960]
-
blk.27.attn_k.weightF16[960 320]
-
blk.27.attn_output.weightF16[960 960]
-
blk.27.attn_q.weightF16[960 960]
-
blk.27.attn_v.weightF16[960 320]
-
blk.28.attn_norm.weightF32[960]
-
blk.28.ffn_down.weightF16[2560 960]
-
blk.28.ffn_gate.weightF16[960 2560]
-
blk.28.ffn_up.weightF16[960 2560]
-
blk.28.ffn_norm.weightF32[960]
-
blk.28.attn_k.weightF16[960 320]
-
blk.28.attn_output.weightF16[960 960]
-
blk.28.attn_q.weightF16[960 960]
-
blk.28.attn_v.weightF16[960 320]
-
blk.29.attn_norm.weightF32[960]
-
blk.29.ffn_down.weightF16[2560 960]
-
blk.29.ffn_gate.weightF16[960 2560]
-
blk.29.ffn_up.weightF16[960 2560]
-
blk.29.ffn_norm.weightF32[960]
-
blk.29.attn_k.weightF16[960 320]
-
blk.29.attn_output.weightF16[960 960]
-
blk.29.attn_q.weightF16[960 960]
-
blk.29.attn_v.weightF16[960 320]
-
blk.30.attn_norm.weightF32[960]
-
blk.30.ffn_down.weightF16[2560 960]
-
blk.30.ffn_gate.weightF16[960 2560]
-
blk.30.ffn_up.weightF16[960 2560]
-
blk.30.ffn_norm.weightF32[960]
-
blk.30.attn_k.weightF16[960 320]
-
blk.30.attn_output.weightF16[960 960]
-
blk.30.attn_q.weightF16[960 960]
-
blk.30.attn_v.weightF16[960 320]
-
blk.31.attn_norm.weightF32[960]
-
blk.31.ffn_down.weightF16[2560 960]
-
blk.31.ffn_gate.weightF16[960 2560]
-
blk.31.ffn_up.weightF16[960 2560]
-
blk.31.ffn_norm.weightF32[960]
-
blk.31.attn_k.weightF16[960 320]
-
blk.31.attn_output.weightF16[960 960]
-
blk.31.attn_q.weightF16[960 960]
-
blk.31.attn_v.weightF16[960 320]
-
output_norm.weightF32[960]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31