latest
5.1GB
Mistral code finetune for ollama
7B
79 Pulls Updated 4 months ago
93a8b51c264d · 5.1GB
-
general.architecturellama
-
general.file_typeQ5_K_M
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length32768
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base10000
-
llama.vocab_size32000
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id2
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.padding_token_id0
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 6 6 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
tokenizer.ggml.unknown_token_id0
-
NameTypeShape
-
token_embd.weightQ5_K[4096 32000]
-
blk.0.attn_q.weightQ5_K[4096 4096]
-
blk.0.attn_k.weightQ5_K[4096 1024]
-
blk.0.attn_v.weightQ6_K[4096 1024]
-
blk.0.attn_output.weightQ5_K[4096 4096]
-
blk.0.ffn_gate.weightQ5_K[4096 14336]
-
blk.0.ffn_up.weightQ5_K[4096 14336]
-
blk.0.ffn_down.weightQ6_K[14336 4096]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.1.attn_q.weightQ5_K[4096 4096]
-
blk.1.attn_k.weightQ5_K[4096 1024]
-
blk.1.attn_v.weightQ6_K[4096 1024]
-
blk.1.attn_output.weightQ5_K[4096 4096]
-
blk.1.ffn_gate.weightQ5_K[4096 14336]
-
blk.1.ffn_up.weightQ5_K[4096 14336]
-
blk.1.ffn_down.weightQ6_K[14336 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.2.attn_q.weightQ5_K[4096 4096]
-
blk.2.attn_k.weightQ5_K[4096 1024]
-
blk.2.attn_v.weightQ6_K[4096 1024]
-
blk.2.attn_output.weightQ5_K[4096 4096]
-
blk.2.ffn_gate.weightQ5_K[4096 14336]
-
blk.2.ffn_up.weightQ5_K[4096 14336]
-
blk.2.ffn_down.weightQ6_K[14336 4096]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.3.attn_q.weightQ5_K[4096 4096]
-
blk.3.attn_k.weightQ5_K[4096 1024]
-
blk.3.attn_v.weightQ6_K[4096 1024]
-
blk.3.attn_output.weightQ5_K[4096 4096]
-
blk.3.ffn_gate.weightQ5_K[4096 14336]
-
blk.3.ffn_up.weightQ5_K[4096 14336]
-
blk.3.ffn_down.weightQ6_K[14336 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.4.attn_q.weightQ5_K[4096 4096]
-
blk.4.attn_k.weightQ5_K[4096 1024]
-
blk.4.attn_v.weightQ5_K[4096 1024]
-
blk.4.attn_output.weightQ5_K[4096 4096]
-
blk.4.ffn_gate.weightQ5_K[4096 14336]
-
blk.4.ffn_up.weightQ5_K[4096 14336]
-
blk.4.ffn_down.weightQ5_K[14336 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.5.attn_q.weightQ5_K[4096 4096]
-
blk.5.attn_k.weightQ5_K[4096 1024]
-
blk.5.attn_v.weightQ5_K[4096 1024]
-
blk.5.attn_output.weightQ5_K[4096 4096]
-
blk.5.ffn_gate.weightQ5_K[4096 14336]
-
blk.5.ffn_up.weightQ5_K[4096 14336]
-
blk.5.ffn_down.weightQ5_K[14336 4096]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.6.attn_q.weightQ5_K[4096 4096]
-
blk.6.attn_k.weightQ5_K[4096 1024]
-
blk.6.attn_v.weightQ6_K[4096 1024]
-
blk.6.attn_output.weightQ5_K[4096 4096]
-
blk.6.ffn_gate.weightQ5_K[4096 14336]
-
blk.6.ffn_up.weightQ5_K[4096 14336]
-
blk.6.ffn_down.weightQ6_K[14336 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.7.attn_q.weightQ5_K[4096 4096]
-
blk.7.attn_k.weightQ5_K[4096 1024]
-
blk.7.attn_v.weightQ5_K[4096 1024]
-
blk.7.attn_output.weightQ5_K[4096 4096]
-
blk.7.ffn_gate.weightQ5_K[4096 14336]
-
blk.7.ffn_up.weightQ5_K[4096 14336]
-
blk.7.ffn_down.weightQ5_K[14336 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.8.attn_q.weightQ5_K[4096 4096]
-
blk.8.attn_k.weightQ5_K[4096 1024]
-
blk.8.attn_v.weightQ5_K[4096 1024]
-
blk.8.attn_output.weightQ5_K[4096 4096]
-
blk.8.ffn_gate.weightQ5_K[4096 14336]
-
blk.8.ffn_up.weightQ5_K[4096 14336]
-
blk.8.ffn_down.weightQ5_K[14336 4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.9.attn_q.weightQ5_K[4096 4096]
-
blk.9.attn_k.weightQ5_K[4096 1024]
-
blk.9.attn_v.weightQ6_K[4096 1024]
-
blk.9.attn_output.weightQ5_K[4096 4096]
-
blk.9.ffn_gate.weightQ5_K[4096 14336]
-
blk.9.ffn_up.weightQ5_K[4096 14336]
-
blk.9.ffn_down.weightQ6_K[14336 4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.10.attn_q.weightQ5_K[4096 4096]
-
blk.10.attn_k.weightQ5_K[4096 1024]
-
blk.10.attn_v.weightQ5_K[4096 1024]
-
blk.10.attn_output.weightQ5_K[4096 4096]
-
blk.10.ffn_gate.weightQ5_K[4096 14336]
-
blk.10.ffn_up.weightQ5_K[4096 14336]
-
blk.10.ffn_down.weightQ5_K[14336 4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.11.attn_q.weightQ5_K[4096 4096]
-
blk.11.attn_k.weightQ5_K[4096 1024]
-
blk.11.attn_v.weightQ5_K[4096 1024]
-
blk.11.attn_output.weightQ5_K[4096 4096]
-
blk.11.ffn_gate.weightQ5_K[4096 14336]
-
blk.11.ffn_up.weightQ5_K[4096 14336]
-
blk.11.ffn_down.weightQ5_K[14336 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.12.attn_q.weightQ5_K[4096 4096]
-
blk.12.attn_k.weightQ5_K[4096 1024]
-
blk.12.attn_v.weightQ6_K[4096 1024]
-
blk.12.attn_output.weightQ5_K[4096 4096]
-
blk.12.ffn_gate.weightQ5_K[4096 14336]
-
blk.12.ffn_up.weightQ5_K[4096 14336]
-
blk.12.ffn_down.weightQ6_K[14336 4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.13.attn_q.weightQ5_K[4096 4096]
-
blk.13.attn_k.weightQ5_K[4096 1024]
-
blk.13.attn_v.weightQ5_K[4096 1024]
-
blk.13.attn_output.weightQ5_K[4096 4096]
-
blk.13.ffn_gate.weightQ5_K[4096 14336]
-
blk.13.ffn_up.weightQ5_K[4096 14336]
-
blk.13.ffn_down.weightQ5_K[14336 4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.14.attn_q.weightQ5_K[4096 4096]
-
blk.14.attn_k.weightQ5_K[4096 1024]
-
blk.14.attn_v.weightQ5_K[4096 1024]
-
blk.14.attn_output.weightQ5_K[4096 4096]
-
blk.14.ffn_gate.weightQ5_K[4096 14336]
-
blk.14.ffn_up.weightQ5_K[4096 14336]
-
blk.14.ffn_down.weightQ5_K[14336 4096]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.15.attn_q.weightQ5_K[4096 4096]
-
blk.15.attn_k.weightQ5_K[4096 1024]
-
blk.15.attn_v.weightQ6_K[4096 1024]
-
blk.15.attn_output.weightQ5_K[4096 4096]
-
blk.15.ffn_gate.weightQ5_K[4096 14336]
-
blk.15.ffn_up.weightQ5_K[4096 14336]
-
blk.15.ffn_down.weightQ6_K[14336 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.16.attn_q.weightQ5_K[4096 4096]
-
blk.16.attn_k.weightQ5_K[4096 1024]
-
blk.16.attn_v.weightQ5_K[4096 1024]
-
blk.16.attn_output.weightQ5_K[4096 4096]
-
blk.16.ffn_gate.weightQ5_K[4096 14336]
-
blk.16.ffn_up.weightQ5_K[4096 14336]
-
blk.16.ffn_down.weightQ5_K[14336 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.17.attn_q.weightQ5_K[4096 4096]
-
blk.17.attn_k.weightQ5_K[4096 1024]
-
blk.17.attn_v.weightQ5_K[4096 1024]
-
blk.17.attn_output.weightQ5_K[4096 4096]
-
blk.17.ffn_gate.weightQ5_K[4096 14336]
-
blk.17.ffn_up.weightQ5_K[4096 14336]
-
blk.17.ffn_down.weightQ5_K[14336 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.18.attn_q.weightQ5_K[4096 4096]
-
blk.18.attn_k.weightQ5_K[4096 1024]
-
blk.18.attn_v.weightQ6_K[4096 1024]
-
blk.18.attn_output.weightQ5_K[4096 4096]
-
blk.18.ffn_gate.weightQ5_K[4096 14336]
-
blk.18.ffn_up.weightQ5_K[4096 14336]
-
blk.18.ffn_down.weightQ6_K[14336 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.19.attn_q.weightQ5_K[4096 4096]
-
blk.19.attn_k.weightQ5_K[4096 1024]
-
blk.19.attn_v.weightQ5_K[4096 1024]
-
blk.19.attn_output.weightQ5_K[4096 4096]
-
blk.19.ffn_gate.weightQ5_K[4096 14336]
-
blk.19.ffn_up.weightQ5_K[4096 14336]
-
blk.19.ffn_down.weightQ5_K[14336 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.20.attn_q.weightQ5_K[4096 4096]
-
blk.20.attn_k.weightQ5_K[4096 1024]
-
blk.20.attn_v.weightQ5_K[4096 1024]
-
blk.20.attn_output.weightQ5_K[4096 4096]
-
blk.20.ffn_gate.weightQ5_K[4096 14336]
-
blk.20.ffn_up.weightQ5_K[4096 14336]
-
blk.20.ffn_down.weightQ5_K[14336 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.21.attn_q.weightQ5_K[4096 4096]
-
blk.21.attn_k.weightQ5_K[4096 1024]
-
blk.21.attn_v.weightQ6_K[4096 1024]
-
blk.21.attn_output.weightQ5_K[4096 4096]
-
blk.21.ffn_gate.weightQ5_K[4096 14336]
-
blk.21.ffn_up.weightQ5_K[4096 14336]
-
blk.21.ffn_down.weightQ6_K[14336 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.22.attn_q.weightQ5_K[4096 4096]
-
blk.22.attn_k.weightQ5_K[4096 1024]
-
blk.22.attn_v.weightQ5_K[4096 1024]
-
blk.22.attn_output.weightQ5_K[4096 4096]
-
blk.22.ffn_gate.weightQ5_K[4096 14336]
-
blk.22.ffn_up.weightQ5_K[4096 14336]
-
blk.22.ffn_down.weightQ5_K[14336 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.23.attn_q.weightQ5_K[4096 4096]
-
blk.23.attn_k.weightQ5_K[4096 1024]
-
blk.23.attn_v.weightQ5_K[4096 1024]
-
blk.23.attn_output.weightQ5_K[4096 4096]
-
blk.23.ffn_gate.weightQ5_K[4096 14336]
-
blk.23.ffn_up.weightQ5_K[4096 14336]
-
blk.23.ffn_down.weightQ5_K[14336 4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.24.attn_q.weightQ5_K[4096 4096]
-
blk.24.attn_k.weightQ5_K[4096 1024]
-
blk.24.attn_v.weightQ6_K[4096 1024]
-
blk.24.attn_output.weightQ5_K[4096 4096]
-
blk.24.ffn_gate.weightQ5_K[4096 14336]
-
blk.24.ffn_up.weightQ5_K[4096 14336]
-
blk.24.ffn_down.weightQ6_K[14336 4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.25.attn_q.weightQ5_K[4096 4096]
-
blk.25.attn_k.weightQ5_K[4096 1024]
-
blk.25.attn_v.weightQ5_K[4096 1024]
-
blk.25.attn_output.weightQ5_K[4096 4096]
-
blk.25.ffn_gate.weightQ5_K[4096 14336]
-
blk.25.ffn_up.weightQ5_K[4096 14336]
-
blk.25.ffn_down.weightQ5_K[14336 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.26.attn_q.weightQ5_K[4096 4096]
-
blk.26.attn_k.weightQ5_K[4096 1024]
-
blk.26.attn_v.weightQ5_K[4096 1024]
-
blk.26.attn_output.weightQ5_K[4096 4096]
-
blk.26.ffn_gate.weightQ5_K[4096 14336]
-
blk.26.ffn_up.weightQ5_K[4096 14336]
-
blk.26.ffn_down.weightQ5_K[14336 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.27.attn_q.weightQ5_K[4096 4096]
-
blk.27.attn_k.weightQ5_K[4096 1024]
-
blk.27.attn_v.weightQ6_K[4096 1024]
-
blk.27.attn_output.weightQ5_K[4096 4096]
-
blk.27.ffn_gate.weightQ5_K[4096 14336]
-
blk.27.ffn_up.weightQ5_K[4096 14336]
-
blk.27.ffn_down.weightQ6_K[14336 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.28.attn_q.weightQ5_K[4096 4096]
-
blk.28.attn_k.weightQ5_K[4096 1024]
-
blk.28.attn_v.weightQ6_K[4096 1024]
-
blk.28.attn_output.weightQ5_K[4096 4096]
-
blk.28.ffn_gate.weightQ5_K[4096 14336]
-
blk.28.ffn_up.weightQ5_K[4096 14336]
-
blk.28.ffn_down.weightQ6_K[14336 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.29.attn_q.weightQ5_K[4096 4096]
-
blk.29.attn_k.weightQ5_K[4096 1024]
-
blk.29.attn_v.weightQ6_K[4096 1024]
-
blk.29.attn_output.weightQ5_K[4096 4096]
-
blk.29.ffn_gate.weightQ5_K[4096 14336]
-
blk.29.ffn_up.weightQ5_K[4096 14336]
-
blk.29.ffn_down.weightQ6_K[14336 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.30.attn_q.weightQ5_K[4096 4096]
-
blk.30.attn_k.weightQ5_K[4096 1024]
-
blk.30.attn_v.weightQ6_K[4096 1024]
-
blk.30.attn_output.weightQ5_K[4096 4096]
-
blk.30.ffn_gate.weightQ5_K[4096 14336]
-
blk.30.ffn_up.weightQ5_K[4096 14336]
-
blk.30.ffn_down.weightQ6_K[14336 4096]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.31.attn_q.weightQ5_K[4096 4096]
-
blk.31.attn_k.weightQ5_K[4096 1024]
-
blk.31.attn_v.weightQ6_K[4096 1024]
-
blk.31.attn_output.weightQ5_K[4096 4096]
-
blk.31.ffn_gate.weightQ5_K[4096 14336]
-
blk.31.ffn_up.weightQ5_K[4096 14336]
-
blk.31.ffn_down.weightQ6_K[14336 4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_norm.weightF32[4096]
-
output_norm.weightF32[4096]
-
output.weightQ6_K[4096 32000]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31