latest
4.9GB
This model is used to translate modern Chinese into Classical Chinese. The dataset used is "The Seventy Biographies"(七十列传) from "Records of the Grand Historian"(史记).
8B
3 Pulls Updated 4 days ago
60bad3551916 · 4.9GB
-
general.architecturellama
-
general.file_typeQ4_K_M
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length131072
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base500000
-
llama.vocab_size128256
-
tokenizer.ggml.bos_token_id128000
-
tokenizer.ggml.eos_token_id128009
-
tokenizer.ggml.merges[Ġ Ġ Ġ ĠĠĠ ĠĠ ĠĠ ĠĠĠ Ġ i n ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.padding_token_id128004
-
tokenizer.ggml.prellama-bpe
-
tokenizer.ggml.token_type[1 1 1 1 1 ...]
-
tokenizer.ggml.tokens[! " # $ % ...]
-
NameTypeShape
-
token_embd.weightQ4_K[4096 128256]
-
blk.0.attn_q.weightQ4_K[4096 4096]
-
blk.0.attn_k.weightQ4_K[4096 1024]
-
blk.0.attn_v.weightQ6_K[4096 1024]
-
blk.0.attn_output.weightQ4_K[4096 4096]
-
blk.0.ffn_gate.weightQ4_K[4096 14336]
-
blk.0.ffn_up.weightQ4_K[4096 14336]
-
blk.0.ffn_down.weightQ6_K[14336 4096]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.1.attn_q.weightQ4_K[4096 4096]
-
blk.1.attn_k.weightQ4_K[4096 1024]
-
blk.1.attn_v.weightQ6_K[4096 1024]
-
blk.1.attn_output.weightQ4_K[4096 4096]
-
blk.1.ffn_gate.weightQ4_K[4096 14336]
-
blk.1.ffn_up.weightQ4_K[4096 14336]
-
blk.1.ffn_down.weightQ6_K[14336 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.2.attn_q.weightQ4_K[4096 4096]
-
blk.2.attn_k.weightQ4_K[4096 1024]
-
blk.2.attn_v.weightQ6_K[4096 1024]
-
blk.2.attn_output.weightQ4_K[4096 4096]
-
blk.2.ffn_gate.weightQ4_K[4096 14336]
-
blk.2.ffn_up.weightQ4_K[4096 14336]
-
blk.2.ffn_down.weightQ6_K[14336 4096]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.3.attn_q.weightQ4_K[4096 4096]
-
blk.3.attn_k.weightQ4_K[4096 1024]
-
blk.3.attn_v.weightQ6_K[4096 1024]
-
blk.3.attn_output.weightQ4_K[4096 4096]
-
blk.3.ffn_gate.weightQ4_K[4096 14336]
-
blk.3.ffn_up.weightQ4_K[4096 14336]
-
blk.3.ffn_down.weightQ6_K[14336 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.4.attn_q.weightQ4_K[4096 4096]
-
blk.4.attn_k.weightQ4_K[4096 1024]
-
blk.4.attn_v.weightQ4_K[4096 1024]
-
blk.4.attn_output.weightQ4_K[4096 4096]
-
blk.4.ffn_gate.weightQ4_K[4096 14336]
-
blk.4.ffn_up.weightQ4_K[4096 14336]
-
blk.4.ffn_down.weightQ4_K[14336 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.5.attn_q.weightQ4_K[4096 4096]
-
blk.5.attn_k.weightQ4_K[4096 1024]
-
blk.5.attn_v.weightQ4_K[4096 1024]
-
blk.5.attn_output.weightQ4_K[4096 4096]
-
blk.5.ffn_gate.weightQ4_K[4096 14336]
-
blk.5.ffn_up.weightQ4_K[4096 14336]
-
blk.5.ffn_down.weightQ4_K[14336 4096]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.6.attn_q.weightQ4_K[4096 4096]
-
blk.6.attn_k.weightQ4_K[4096 1024]
-
blk.6.attn_v.weightQ6_K[4096 1024]
-
blk.6.attn_output.weightQ4_K[4096 4096]
-
blk.6.ffn_gate.weightQ4_K[4096 14336]
-
blk.6.ffn_up.weightQ4_K[4096 14336]
-
blk.6.ffn_down.weightQ6_K[14336 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.7.attn_q.weightQ4_K[4096 4096]
-
blk.7.attn_k.weightQ4_K[4096 1024]
-
blk.7.attn_v.weightQ4_K[4096 1024]
-
blk.7.attn_output.weightQ4_K[4096 4096]
-
blk.7.ffn_gate.weightQ4_K[4096 14336]
-
blk.7.ffn_up.weightQ4_K[4096 14336]
-
blk.7.ffn_down.weightQ4_K[14336 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.8.attn_q.weightQ4_K[4096 4096]
-
blk.8.attn_k.weightQ4_K[4096 1024]
-
blk.8.attn_v.weightQ4_K[4096 1024]
-
blk.8.attn_output.weightQ4_K[4096 4096]
-
blk.8.ffn_gate.weightQ4_K[4096 14336]
-
blk.8.ffn_up.weightQ4_K[4096 14336]
-
blk.8.ffn_down.weightQ4_K[14336 4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.9.attn_q.weightQ4_K[4096 4096]
-
blk.9.attn_k.weightQ4_K[4096 1024]
-
blk.9.attn_v.weightQ6_K[4096 1024]
-
blk.9.attn_output.weightQ4_K[4096 4096]
-
blk.9.ffn_gate.weightQ4_K[4096 14336]
-
blk.9.ffn_up.weightQ4_K[4096 14336]
-
blk.9.ffn_down.weightQ6_K[14336 4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.10.attn_q.weightQ4_K[4096 4096]
-
blk.10.attn_k.weightQ4_K[4096 1024]
-
blk.10.attn_v.weightQ4_K[4096 1024]
-
blk.10.attn_output.weightQ4_K[4096 4096]
-
blk.10.ffn_gate.weightQ4_K[4096 14336]
-
blk.10.ffn_up.weightQ4_K[4096 14336]
-
blk.10.ffn_down.weightQ4_K[14336 4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.11.attn_q.weightQ4_K[4096 4096]
-
blk.11.attn_k.weightQ4_K[4096 1024]
-
blk.11.attn_v.weightQ4_K[4096 1024]
-
blk.11.attn_output.weightQ4_K[4096 4096]
-
blk.11.ffn_gate.weightQ4_K[4096 14336]
-
blk.11.ffn_up.weightQ4_K[4096 14336]
-
blk.11.ffn_down.weightQ4_K[14336 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.12.attn_q.weightQ4_K[4096 4096]
-
blk.12.attn_k.weightQ4_K[4096 1024]
-
blk.12.attn_v.weightQ6_K[4096 1024]
-
blk.12.attn_output.weightQ4_K[4096 4096]
-
blk.12.ffn_gate.weightQ4_K[4096 14336]
-
blk.12.ffn_up.weightQ4_K[4096 14336]
-
blk.12.ffn_down.weightQ6_K[14336 4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.13.attn_q.weightQ4_K[4096 4096]
-
blk.13.attn_k.weightQ4_K[4096 1024]
-
blk.13.attn_v.weightQ4_K[4096 1024]
-
blk.13.attn_output.weightQ4_K[4096 4096]
-
blk.13.ffn_gate.weightQ4_K[4096 14336]
-
blk.13.ffn_up.weightQ4_K[4096 14336]
-
blk.13.ffn_down.weightQ4_K[14336 4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.14.attn_q.weightQ4_K[4096 4096]
-
blk.14.attn_k.weightQ4_K[4096 1024]
-
blk.14.attn_v.weightQ4_K[4096 1024]
-
blk.14.attn_output.weightQ4_K[4096 4096]
-
blk.14.ffn_gate.weightQ4_K[4096 14336]
-
blk.14.ffn_up.weightQ4_K[4096 14336]
-
blk.14.ffn_down.weightQ4_K[14336 4096]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.15.attn_q.weightQ4_K[4096 4096]
-
blk.15.attn_k.weightQ4_K[4096 1024]
-
blk.15.attn_v.weightQ6_K[4096 1024]
-
blk.15.attn_output.weightQ4_K[4096 4096]
-
blk.15.ffn_gate.weightQ4_K[4096 14336]
-
blk.15.ffn_up.weightQ4_K[4096 14336]
-
blk.15.ffn_down.weightQ6_K[14336 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.16.attn_q.weightQ4_K[4096 4096]
-
blk.16.attn_k.weightQ4_K[4096 1024]
-
blk.16.attn_v.weightQ4_K[4096 1024]
-
blk.16.attn_output.weightQ4_K[4096 4096]
-
blk.16.ffn_gate.weightQ4_K[4096 14336]
-
blk.16.ffn_up.weightQ4_K[4096 14336]
-
blk.16.ffn_down.weightQ4_K[14336 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.17.attn_q.weightQ4_K[4096 4096]
-
blk.17.attn_k.weightQ4_K[4096 1024]
-
blk.17.attn_v.weightQ4_K[4096 1024]
-
blk.17.attn_output.weightQ4_K[4096 4096]
-
blk.17.ffn_gate.weightQ4_K[4096 14336]
-
blk.17.ffn_up.weightQ4_K[4096 14336]
-
blk.17.ffn_down.weightQ4_K[14336 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.18.attn_q.weightQ4_K[4096 4096]
-
blk.18.attn_k.weightQ4_K[4096 1024]
-
blk.18.attn_v.weightQ6_K[4096 1024]
-
blk.18.attn_output.weightQ4_K[4096 4096]
-
blk.18.ffn_gate.weightQ4_K[4096 14336]
-
blk.18.ffn_up.weightQ4_K[4096 14336]
-
blk.18.ffn_down.weightQ6_K[14336 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.19.attn_q.weightQ4_K[4096 4096]
-
blk.19.attn_k.weightQ4_K[4096 1024]
-
blk.19.attn_v.weightQ4_K[4096 1024]
-
blk.19.attn_output.weightQ4_K[4096 4096]
-
blk.19.ffn_gate.weightQ4_K[4096 14336]
-
blk.19.ffn_up.weightQ4_K[4096 14336]
-
blk.19.ffn_down.weightQ4_K[14336 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.20.attn_q.weightQ4_K[4096 4096]
-
blk.20.attn_k.weightQ4_K[4096 1024]
-
blk.20.attn_v.weightQ4_K[4096 1024]
-
blk.20.attn_output.weightQ4_K[4096 4096]
-
blk.20.ffn_gate.weightQ4_K[4096 14336]
-
blk.20.ffn_up.weightQ4_K[4096 14336]
-
blk.20.ffn_down.weightQ4_K[14336 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.21.attn_q.weightQ4_K[4096 4096]
-
blk.21.attn_k.weightQ4_K[4096 1024]
-
blk.21.attn_v.weightQ6_K[4096 1024]
-
blk.21.attn_output.weightQ4_K[4096 4096]
-
blk.21.ffn_gate.weightQ4_K[4096 14336]
-
blk.21.ffn_up.weightQ4_K[4096 14336]
-
blk.21.ffn_down.weightQ6_K[14336 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.22.attn_q.weightQ4_K[4096 4096]
-
blk.22.attn_k.weightQ4_K[4096 1024]
-
blk.22.attn_v.weightQ4_K[4096 1024]
-
blk.22.attn_output.weightQ4_K[4096 4096]
-
blk.22.ffn_gate.weightQ4_K[4096 14336]
-
blk.22.ffn_up.weightQ4_K[4096 14336]
-
blk.22.ffn_down.weightQ4_K[14336 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.23.attn_q.weightQ4_K[4096 4096]
-
blk.23.attn_k.weightQ4_K[4096 1024]
-
blk.23.attn_v.weightQ4_K[4096 1024]
-
blk.23.attn_output.weightQ4_K[4096 4096]
-
blk.23.ffn_gate.weightQ4_K[4096 14336]
-
blk.23.ffn_up.weightQ4_K[4096 14336]
-
blk.23.ffn_down.weightQ4_K[14336 4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.24.attn_q.weightQ4_K[4096 4096]
-
blk.24.attn_k.weightQ4_K[4096 1024]
-
blk.24.attn_v.weightQ6_K[4096 1024]
-
blk.24.attn_output.weightQ4_K[4096 4096]
-
blk.24.ffn_gate.weightQ4_K[4096 14336]
-
blk.24.ffn_up.weightQ4_K[4096 14336]
-
blk.24.ffn_down.weightQ6_K[14336 4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.25.attn_q.weightQ4_K[4096 4096]
-
blk.25.attn_k.weightQ4_K[4096 1024]
-
blk.25.attn_v.weightQ4_K[4096 1024]
-
blk.25.attn_output.weightQ4_K[4096 4096]
-
blk.25.ffn_gate.weightQ4_K[4096 14336]
-
blk.25.ffn_up.weightQ4_K[4096 14336]
-
blk.25.ffn_down.weightQ4_K[14336 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.26.attn_q.weightQ4_K[4096 4096]
-
blk.26.attn_k.weightQ4_K[4096 1024]
-
blk.26.attn_v.weightQ4_K[4096 1024]
-
blk.26.attn_output.weightQ4_K[4096 4096]
-
blk.26.ffn_gate.weightQ4_K[4096 14336]
-
blk.26.ffn_up.weightQ4_K[4096 14336]
-
blk.26.ffn_down.weightQ4_K[14336 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.27.attn_q.weightQ4_K[4096 4096]
-
blk.27.attn_k.weightQ4_K[4096 1024]
-
blk.27.attn_v.weightQ6_K[4096 1024]
-
blk.27.attn_output.weightQ4_K[4096 4096]
-
blk.27.ffn_gate.weightQ4_K[4096 14336]
-
blk.27.ffn_up.weightQ4_K[4096 14336]
-
blk.27.ffn_down.weightQ6_K[14336 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.28.attn_q.weightQ4_K[4096 4096]
-
blk.28.attn_k.weightQ4_K[4096 1024]
-
blk.28.attn_v.weightQ6_K[4096 1024]
-
blk.28.attn_output.weightQ4_K[4096 4096]
-
blk.28.ffn_gate.weightQ4_K[4096 14336]
-
blk.28.ffn_up.weightQ4_K[4096 14336]
-
blk.28.ffn_down.weightQ6_K[14336 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.29.attn_q.weightQ4_K[4096 4096]
-
blk.29.attn_k.weightQ4_K[4096 1024]
-
blk.29.attn_v.weightQ6_K[4096 1024]
-
blk.29.attn_output.weightQ4_K[4096 4096]
-
blk.29.ffn_gate.weightQ4_K[4096 14336]
-
blk.29.ffn_up.weightQ4_K[4096 14336]
-
blk.29.ffn_down.weightQ6_K[14336 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.30.attn_q.weightQ4_K[4096 4096]
-
blk.30.attn_k.weightQ4_K[4096 1024]
-
blk.30.attn_v.weightQ6_K[4096 1024]
-
blk.30.attn_output.weightQ4_K[4096 4096]
-
blk.30.ffn_gate.weightQ4_K[4096 14336]
-
blk.30.ffn_up.weightQ4_K[4096 14336]
-
blk.30.ffn_down.weightQ6_K[14336 4096]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.31.attn_q.weightQ4_K[4096 4096]
-
blk.31.attn_k.weightQ4_K[4096 1024]
-
blk.31.attn_v.weightQ6_K[4096 1024]
-
blk.31.attn_output.weightQ4_K[4096 4096]
-
blk.31.ffn_gate.weightQ4_K[4096 14336]
-
blk.31.ffn_up.weightQ4_K[4096 14336]
-
blk.31.ffn_down.weightQ6_K[14336 4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_norm.weightF32[4096]
-
output_norm.weightF32[4096]
-
output.weightQ6_K[4096 128256]
-
rope_freqs.weightF32[64]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31