latest
4.4GB
Italian mistral finetuning
13 Pulls Updated 3 months ago
678d208a8993 · 4.4GB
-
general.architecturellama
-
general.file_typeQ4_K_M
-
general.quantization_version2
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length32768
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base10000
-
llama.vocab_size32002
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id32000
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.padding_token_id2
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 6 6 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
tokenizer.ggml.unknown_token_id0
-
NameTypeShape
-
token_embd.weightQ4_K[4096 32002]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightQ6_K[14336 4096]
-
blk.0.ffn_gate.weightQ4_K[4096 14336]
-
blk.0.ffn_up.weightQ4_K[4096 14336]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightQ4_K[4096 1024]
-
blk.0.attn_output.weightQ4_K[4096 4096]
-
blk.0.attn_q.weightQ4_K[4096 4096]
-
blk.0.attn_v.weightQ6_K[4096 1024]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightQ6_K[14336 4096]
-
blk.1.ffn_gate.weightQ4_K[4096 14336]
-
blk.1.ffn_up.weightQ4_K[4096 14336]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightQ4_K[4096 1024]
-
blk.1.attn_output.weightQ4_K[4096 4096]
-
blk.1.attn_q.weightQ4_K[4096 4096]
-
blk.1.attn_v.weightQ6_K[4096 1024]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightQ6_K[14336 4096]
-
blk.2.ffn_gate.weightQ4_K[4096 14336]
-
blk.2.ffn_up.weightQ4_K[4096 14336]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightQ4_K[4096 1024]
-
blk.2.attn_output.weightQ4_K[4096 4096]
-
blk.2.attn_q.weightQ4_K[4096 4096]
-
blk.2.attn_v.weightQ6_K[4096 1024]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightQ6_K[14336 4096]
-
blk.3.ffn_gate.weightQ4_K[4096 14336]
-
blk.3.ffn_up.weightQ4_K[4096 14336]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightQ4_K[4096 1024]
-
blk.3.attn_output.weightQ4_K[4096 4096]
-
blk.3.attn_q.weightQ4_K[4096 4096]
-
blk.3.attn_v.weightQ4_K[4096 1024]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightQ4_K[14336 4096]
-
blk.4.ffn_gate.weightQ4_K[4096 14336]
-
blk.4.ffn_up.weightQ4_K[4096 14336]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightQ4_K[4096 1024]
-
blk.4.attn_output.weightQ4_K[4096 4096]
-
blk.4.attn_q.weightQ4_K[4096 4096]
-
blk.4.attn_v.weightQ4_K[4096 1024]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightQ4_K[14336 4096]
-
blk.5.ffn_gate.weightQ4_K[4096 14336]
-
blk.5.ffn_up.weightQ4_K[4096 14336]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightQ4_K[4096 1024]
-
blk.5.attn_output.weightQ4_K[4096 4096]
-
blk.5.attn_q.weightQ4_K[4096 4096]
-
blk.5.attn_v.weightQ6_K[4096 1024]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightQ6_K[14336 4096]
-
blk.6.ffn_gate.weightQ4_K[4096 14336]
-
blk.6.ffn_up.weightQ4_K[4096 14336]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightQ4_K[4096 1024]
-
blk.6.attn_output.weightQ4_K[4096 4096]
-
blk.6.attn_q.weightQ4_K[4096 4096]
-
blk.6.attn_v.weightQ4_K[4096 1024]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightQ4_K[14336 4096]
-
blk.7.ffn_gate.weightQ4_K[4096 14336]
-
blk.7.ffn_up.weightQ4_K[4096 14336]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightQ4_K[4096 1024]
-
blk.7.attn_output.weightQ4_K[4096 4096]
-
blk.7.attn_q.weightQ4_K[4096 4096]
-
blk.7.attn_v.weightQ4_K[4096 1024]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightQ4_K[14336 4096]
-
blk.8.ffn_gate.weightQ4_K[4096 14336]
-
blk.8.ffn_up.weightQ4_K[4096 14336]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightQ4_K[4096 1024]
-
blk.8.attn_output.weightQ4_K[4096 4096]
-
blk.8.attn_q.weightQ4_K[4096 4096]
-
blk.8.attn_v.weightQ6_K[4096 1024]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightQ6_K[14336 4096]
-
blk.9.ffn_gate.weightQ4_K[4096 14336]
-
blk.9.ffn_up.weightQ4_K[4096 14336]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightQ4_K[4096 1024]
-
blk.9.attn_output.weightQ4_K[4096 4096]
-
blk.9.attn_q.weightQ4_K[4096 4096]
-
blk.9.attn_v.weightQ4_K[4096 1024]
-
blk.10.ffn_gate.weightQ4_K[4096 14336]
-
blk.10.ffn_up.weightQ4_K[4096 14336]
-
blk.10.attn_k.weightQ4_K[4096 1024]
-
blk.10.attn_output.weightQ4_K[4096 4096]
-
blk.10.attn_q.weightQ4_K[4096 4096]
-
blk.10.attn_v.weightQ6_K[4096 1024]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightQ4_K[14336 4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightQ4_K[14336 4096]
-
blk.11.ffn_gate.weightQ4_K[4096 14336]
-
blk.11.ffn_up.weightQ4_K[4096 14336]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightQ4_K[4096 1024]
-
blk.11.attn_output.weightQ4_K[4096 4096]
-
blk.11.attn_q.weightQ4_K[4096 4096]
-
blk.11.attn_v.weightQ4_K[4096 1024]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightQ6_K[14336 4096]
-
blk.12.ffn_gate.weightQ4_K[4096 14336]
-
blk.12.ffn_up.weightQ4_K[4096 14336]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightQ4_K[4096 1024]
-
blk.12.attn_output.weightQ4_K[4096 4096]
-
blk.12.attn_q.weightQ4_K[4096 4096]
-
blk.12.attn_v.weightQ6_K[4096 1024]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightQ4_K[14336 4096]
-
blk.13.ffn_gate.weightQ4_K[4096 14336]
-
blk.13.ffn_up.weightQ4_K[4096 14336]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightQ4_K[4096 1024]
-
blk.13.attn_output.weightQ4_K[4096 4096]
-
blk.13.attn_q.weightQ4_K[4096 4096]
-
blk.13.attn_v.weightQ4_K[4096 1024]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightQ4_K[14336 4096]
-
blk.14.ffn_gate.weightQ4_K[4096 14336]
-
blk.14.ffn_up.weightQ4_K[4096 14336]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightQ4_K[4096 1024]
-
blk.14.attn_output.weightQ4_K[4096 4096]
-
blk.14.attn_q.weightQ4_K[4096 4096]
-
blk.14.attn_v.weightQ4_K[4096 1024]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightQ6_K[14336 4096]
-
blk.15.ffn_gate.weightQ4_K[4096 14336]
-
blk.15.ffn_up.weightQ4_K[4096 14336]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightQ4_K[4096 1024]
-
blk.15.attn_output.weightQ4_K[4096 4096]
-
blk.15.attn_q.weightQ4_K[4096 4096]
-
blk.15.attn_v.weightQ6_K[4096 1024]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightQ4_K[14336 4096]
-
blk.16.ffn_gate.weightQ4_K[4096 14336]
-
blk.16.ffn_up.weightQ4_K[4096 14336]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightQ4_K[4096 1024]
-
blk.16.attn_output.weightQ4_K[4096 4096]
-
blk.16.attn_q.weightQ4_K[4096 4096]
-
blk.16.attn_v.weightQ4_K[4096 1024]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightQ4_K[14336 4096]
-
blk.17.ffn_gate.weightQ4_K[4096 14336]
-
blk.17.ffn_up.weightQ4_K[4096 14336]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightQ4_K[4096 1024]
-
blk.17.attn_output.weightQ4_K[4096 4096]
-
blk.17.attn_q.weightQ4_K[4096 4096]
-
blk.17.attn_v.weightQ4_K[4096 1024]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightQ6_K[14336 4096]
-
blk.18.ffn_gate.weightQ4_K[4096 14336]
-
blk.18.ffn_up.weightQ4_K[4096 14336]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightQ4_K[4096 1024]
-
blk.18.attn_output.weightQ4_K[4096 4096]
-
blk.18.attn_q.weightQ4_K[4096 4096]
-
blk.18.attn_v.weightQ6_K[4096 1024]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightQ4_K[14336 4096]
-
blk.19.ffn_gate.weightQ4_K[4096 14336]
-
blk.19.ffn_up.weightQ4_K[4096 14336]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightQ4_K[4096 1024]
-
blk.19.attn_output.weightQ4_K[4096 4096]
-
blk.19.attn_q.weightQ4_K[4096 4096]
-
blk.19.attn_v.weightQ4_K[4096 1024]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightQ4_K[14336 4096]
-
blk.20.ffn_gate.weightQ4_K[4096 14336]
-
blk.20.ffn_up.weightQ4_K[4096 14336]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.weightQ4_K[4096 1024]
-
blk.20.attn_output.weightQ4_K[4096 4096]
-
blk.20.attn_q.weightQ4_K[4096 4096]
-
blk.20.attn_v.weightQ4_K[4096 1024]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightQ6_K[14336 4096]
-
blk.21.ffn_gate.weightQ4_K[4096 14336]
-
blk.21.ffn_up.weightQ4_K[4096 14336]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightQ4_K[4096 1024]
-
blk.21.attn_output.weightQ4_K[4096 4096]
-
blk.21.attn_q.weightQ4_K[4096 4096]
-
blk.21.attn_v.weightQ6_K[4096 1024]
-
blk.22.attn_k.weightQ4_K[4096 1024]
-
blk.22.attn_output.weightQ4_K[4096 4096]
-
blk.22.attn_q.weightQ4_K[4096 4096]
-
blk.22.attn_v.weightQ4_K[4096 1024]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightQ4_K[14336 4096]
-
blk.22.ffn_gate.weightQ4_K[4096 14336]
-
blk.22.ffn_up.weightQ4_K[4096 14336]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightQ4_K[14336 4096]
-
blk.23.ffn_gate.weightQ4_K[4096 14336]
-
blk.23.ffn_up.weightQ4_K[4096 14336]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightQ4_K[4096 1024]
-
blk.23.attn_output.weightQ4_K[4096 4096]
-
blk.23.attn_q.weightQ4_K[4096 4096]
-
blk.23.attn_v.weightQ4_K[4096 1024]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightQ6_K[14336 4096]
-
blk.24.ffn_gate.weightQ4_K[4096 14336]
-
blk.24.ffn_up.weightQ4_K[4096 14336]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightQ4_K[4096 1024]
-
blk.24.attn_output.weightQ4_K[4096 4096]
-
blk.24.attn_q.weightQ4_K[4096 4096]
-
blk.24.attn_v.weightQ6_K[4096 1024]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightQ4_K[14336 4096]
-
blk.25.ffn_gate.weightQ4_K[4096 14336]
-
blk.25.ffn_up.weightQ4_K[4096 14336]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightQ4_K[4096 1024]
-
blk.25.attn_output.weightQ4_K[4096 4096]
-
blk.25.attn_q.weightQ4_K[4096 4096]
-
blk.25.attn_v.weightQ4_K[4096 1024]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightQ4_K[14336 4096]
-
blk.26.ffn_gate.weightQ4_K[4096 14336]
-
blk.26.ffn_up.weightQ4_K[4096 14336]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightQ4_K[4096 1024]
-
blk.26.attn_output.weightQ4_K[4096 4096]
-
blk.26.attn_q.weightQ4_K[4096 4096]
-
blk.26.attn_v.weightQ4_K[4096 1024]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightQ6_K[14336 4096]
-
blk.27.ffn_gate.weightQ4_K[4096 14336]
-
blk.27.ffn_up.weightQ4_K[4096 14336]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightQ4_K[4096 1024]
-
blk.27.attn_output.weightQ4_K[4096 4096]
-
blk.27.attn_q.weightQ4_K[4096 4096]
-
blk.27.attn_v.weightQ6_K[4096 1024]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightQ6_K[14336 4096]
-
blk.28.ffn_gate.weightQ4_K[4096 14336]
-
blk.28.ffn_up.weightQ4_K[4096 14336]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightQ4_K[4096 1024]
-
blk.28.attn_output.weightQ4_K[4096 4096]
-
blk.28.attn_q.weightQ4_K[4096 4096]
-
blk.28.attn_v.weightQ6_K[4096 1024]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightQ6_K[14336 4096]
-
blk.29.ffn_gate.weightQ4_K[4096 14336]
-
blk.29.ffn_up.weightQ4_K[4096 14336]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightQ4_K[4096 1024]
-
blk.29.attn_output.weightQ4_K[4096 4096]
-
blk.29.attn_q.weightQ4_K[4096 4096]
-
blk.29.attn_v.weightQ6_K[4096 1024]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightQ6_K[14336 4096]
-
blk.30.ffn_gate.weightQ4_K[4096 14336]
-
blk.30.ffn_up.weightQ4_K[4096 14336]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightQ4_K[4096 1024]
-
blk.30.attn_output.weightQ4_K[4096 4096]
-
blk.30.attn_q.weightQ4_K[4096 4096]
-
blk.30.attn_v.weightQ6_K[4096 1024]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightQ6_K[14336 4096]
-
blk.31.ffn_gate.weightQ4_K[4096 14336]
-
blk.31.ffn_up.weightQ4_K[4096 14336]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.weightQ4_K[4096 1024]
-
blk.31.attn_output.weightQ4_K[4096 4096]
-
blk.31.attn_q.weightQ4_K[4096 4096]
-
blk.31.attn_v.weightQ6_K[4096 1024]
-
output.weightQ6_K[4096 32002]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31