Quantized version of DeepSeek Coder v1.5 and Q8_0_L quantization of v2 model form bartowski/DeepSeek-Coder-V2-Lite-Base-GGUF and bartowski/DeepSeek-Coder-V2-Lite-Instruct-GGUF
7B
209 Pulls Updated 3 months ago
f4399d415270 · 14GB
-
general.architecturellama
-
general.file_typeF16
-
llama.attention.head_count32
-
llama.attention.head_count_kv32
-
llama.attention.layer_norm_rms_epsilon1e-06
-
llama.block_count30
-
llama.context_length4096
-
llama.embedding_length4096
-
llama.feed_forward_length11008
-
llama.rope.dimension_count128
-
llama.rope.freq_base10000
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id100000
-
tokenizer.ggml.eos_token_id100015
-
tokenizer.ggml.merges[Ġ Ġ Ġ t Ġ a i n h e ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.padding_token_id100001
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[1 1 1 1 1 ...]
-
tokenizer.ggml.tokens[! " # $ % ...]
-
NameTypeShape
-
token_embd.weightF16[4096 102400]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightF16[11008 4096]
-
blk.0.ffn_gate.weightF16[4096 11008]
-
blk.0.ffn_up.weightF16[4096 11008]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightF16[4096 4096]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.attn_q.weightF16[4096 4096]
-
blk.0.attn_v.weightF16[4096 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightF16[11008 4096]
-
blk.1.ffn_gate.weightF16[4096 11008]
-
blk.1.ffn_up.weightF16[4096 11008]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightF16[4096 4096]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.attn_q.weightF16[4096 4096]
-
blk.1.attn_v.weightF16[4096 4096]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightF16[11008 4096]
-
blk.2.ffn_gate.weightF16[4096 11008]
-
blk.2.ffn_up.weightF16[4096 11008]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightF16[4096 4096]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.attn_q.weightF16[4096 4096]
-
blk.2.attn_v.weightF16[4096 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightF16[11008 4096]
-
blk.3.ffn_gate.weightF16[4096 11008]
-
blk.3.ffn_up.weightF16[4096 11008]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightF16[4096 4096]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.attn_q.weightF16[4096 4096]
-
blk.3.attn_v.weightF16[4096 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightF16[11008 4096]
-
blk.4.ffn_gate.weightF16[4096 11008]
-
blk.4.ffn_up.weightF16[4096 11008]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightF16[4096 4096]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.attn_q.weightF16[4096 4096]
-
blk.4.attn_v.weightF16[4096 4096]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightF16[11008 4096]
-
blk.5.ffn_gate.weightF16[4096 11008]
-
blk.5.ffn_up.weightF16[4096 11008]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightF16[4096 4096]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.5.attn_q.weightF16[4096 4096]
-
blk.5.attn_v.weightF16[4096 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightF16[11008 4096]
-
blk.6.ffn_gate.weightF16[4096 11008]
-
blk.6.ffn_up.weightF16[4096 11008]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightF16[4096 4096]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.attn_q.weightF16[4096 4096]
-
blk.6.attn_v.weightF16[4096 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightF16[11008 4096]
-
blk.7.ffn_gate.weightF16[4096 11008]
-
blk.7.ffn_up.weightF16[4096 11008]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightF16[4096 4096]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.attn_q.weightF16[4096 4096]
-
blk.7.attn_v.weightF16[4096 4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightF16[11008 4096]
-
blk.8.ffn_gate.weightF16[4096 11008]
-
blk.8.ffn_up.weightF16[4096 11008]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightF16[4096 4096]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.attn_q.weightF16[4096 4096]
-
blk.8.attn_v.weightF16[4096 4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightF16[11008 4096]
-
blk.9.ffn_gate.weightF16[4096 11008]
-
blk.9.ffn_up.weightF16[4096 11008]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightF16[4096 4096]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.attn_q.weightF16[4096 4096]
-
blk.9.attn_v.weightF16[4096 4096]
-
blk.10.attn_k.weightF16[4096 4096]
-
blk.10.attn_q.weightF16[4096 4096]
-
blk.10.attn_v.weightF16[4096 4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightF16[11008 4096]
-
blk.10.ffn_gate.weightF16[4096 11008]
-
blk.10.ffn_up.weightF16[4096 11008]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightF16[11008 4096]
-
blk.11.ffn_gate.weightF16[4096 11008]
-
blk.11.ffn_up.weightF16[4096 11008]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightF16[4096 4096]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.attn_q.weightF16[4096 4096]
-
blk.11.attn_v.weightF16[4096 4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightF16[11008 4096]
-
blk.12.ffn_gate.weightF16[4096 11008]
-
blk.12.ffn_up.weightF16[4096 11008]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightF16[4096 4096]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.attn_q.weightF16[4096 4096]
-
blk.12.attn_v.weightF16[4096 4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightF16[11008 4096]
-
blk.13.ffn_gate.weightF16[4096 11008]
-
blk.13.ffn_up.weightF16[4096 11008]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightF16[4096 4096]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.attn_q.weightF16[4096 4096]
-
blk.13.attn_v.weightF16[4096 4096]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightF16[11008 4096]
-
blk.14.ffn_gate.weightF16[4096 11008]
-
blk.14.ffn_up.weightF16[4096 11008]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightF16[4096 4096]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.attn_q.weightF16[4096 4096]
-
blk.14.attn_v.weightF16[4096 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightF16[11008 4096]
-
blk.15.ffn_gate.weightF16[4096 11008]
-
blk.15.ffn_up.weightF16[4096 11008]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightF16[4096 4096]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.attn_q.weightF16[4096 4096]
-
blk.15.attn_v.weightF16[4096 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightF16[11008 4096]
-
blk.16.ffn_gate.weightF16[4096 11008]
-
blk.16.ffn_up.weightF16[4096 11008]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightF16[4096 4096]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.attn_q.weightF16[4096 4096]
-
blk.16.attn_v.weightF16[4096 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightF16[11008 4096]
-
blk.17.ffn_gate.weightF16[4096 11008]
-
blk.17.ffn_up.weightF16[4096 11008]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightF16[4096 4096]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.attn_q.weightF16[4096 4096]
-
blk.17.attn_v.weightF16[4096 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightF16[11008 4096]
-
blk.18.ffn_gate.weightF16[4096 11008]
-
blk.18.ffn_up.weightF16[4096 11008]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightF16[4096 4096]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.attn_q.weightF16[4096 4096]
-
blk.18.attn_v.weightF16[4096 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightF16[11008 4096]
-
blk.19.ffn_gate.weightF16[4096 11008]
-
blk.19.ffn_up.weightF16[4096 11008]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightF16[4096 4096]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.attn_q.weightF16[4096 4096]
-
blk.19.attn_v.weightF16[4096 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightF16[11008 4096]
-
blk.20.ffn_gate.weightF16[4096 11008]
-
blk.20.ffn_up.weightF16[4096 11008]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.weightF16[4096 4096]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.attn_q.weightF16[4096 4096]
-
blk.20.attn_v.weightF16[4096 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightF16[11008 4096]
-
blk.21.ffn_gate.weightF16[4096 11008]
-
blk.21.ffn_up.weightF16[4096 11008]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightF16[4096 4096]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.attn_q.weightF16[4096 4096]
-
blk.21.attn_v.weightF16[4096 4096]
-
blk.22.ffn_gate.weightF16[4096 11008]
-
blk.22.attn_k.weightF16[4096 4096]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.attn_q.weightF16[4096 4096]
-
blk.22.attn_v.weightF16[4096 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightF16[11008 4096]
-
blk.22.ffn_up.weightF16[4096 11008]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightF16[11008 4096]
-
blk.23.ffn_gate.weightF16[4096 11008]
-
blk.23.ffn_up.weightF16[4096 11008]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightF16[4096 4096]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.attn_q.weightF16[4096 4096]
-
blk.23.attn_v.weightF16[4096 4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightF16[11008 4096]
-
blk.24.ffn_gate.weightF16[4096 11008]
-
blk.24.ffn_up.weightF16[4096 11008]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightF16[4096 4096]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.attn_q.weightF16[4096 4096]
-
blk.24.attn_v.weightF16[4096 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightF16[11008 4096]
-
blk.25.ffn_gate.weightF16[4096 11008]
-
blk.25.ffn_up.weightF16[4096 11008]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightF16[4096 4096]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.attn_q.weightF16[4096 4096]
-
blk.25.attn_v.weightF16[4096 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightF16[11008 4096]
-
blk.26.ffn_gate.weightF16[4096 11008]
-
blk.26.ffn_up.weightF16[4096 11008]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightF16[4096 4096]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.attn_q.weightF16[4096 4096]
-
blk.26.attn_v.weightF16[4096 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightF16[11008 4096]
-
blk.27.ffn_gate.weightF16[4096 11008]
-
blk.27.ffn_up.weightF16[4096 11008]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightF16[4096 4096]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.27.attn_q.weightF16[4096 4096]
-
blk.27.attn_v.weightF16[4096 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightF16[11008 4096]
-
blk.28.ffn_gate.weightF16[4096 11008]
-
blk.28.ffn_up.weightF16[4096 11008]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightF16[4096 4096]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.attn_q.weightF16[4096 4096]
-
blk.28.attn_v.weightF16[4096 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightF16[11008 4096]
-
blk.29.ffn_gate.weightF16[4096 11008]
-
blk.29.ffn_up.weightF16[4096 11008]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightF16[4096 4096]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.attn_q.weightF16[4096 4096]
-
blk.29.attn_v.weightF16[4096 4096]
-
output.weightF16[4096 102400]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29