latest
1.6MB
1 Pull Updated 2 months ago
6a40c815a788 · 1.6MB
-
general.architecturellama
-
general.file_typeF16
-
llama.attention.head_count32
-
llama.attention.head_count_kv32
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length4096
-
llama.embedding_length4096
-
llama.feed_forward_length11008
-
llama.rope.dimension_count128
-
llama.rope.freq_base0
-
llama.vocab_size32001
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id2
-
tokenizer.ggml.merges[▁ t e r i n ▁ a e n ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.predefault
-
tokenizer.ggml.token_type[3 3 3 1 1 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
tokenizer.ggml.unknown_token_id0
Metadata