Models
GitHub
Discord
Turbo
Sign in
Download
Models
Download
GitHub
Discord
Sign in
takenusername
/
tinyllm-2
:latest
257
Downloads
Updated
8 months ago
This IS the fastest and smartest (probably) llm for the raspberry pi 4b, 5, and newer models if any
This IS the fastest and smartest (probably) llm for the raspberry pi 4b, 5, and newer models if any
Cancel
tinyllm-2:latest
...
/
model
633fc5be925f · 2.2GB
Metadata
general.architecture
phi3
phi3
general.file_type
Q4_0
Q4_0
phi3.attention.head_count
32
32
phi3.attention.head_count_kv
32
32
phi3.attention.layer_norm_rms_epsilon
1e-05
1e-05
phi3.attention.sliding_window
262144
262144
phi3.block_count
32
32
phi3.context_length
131072
131072
phi3.embedding_length
3072
3072
phi3.feed_forward_length
8192
8192
phi3.rope.dimension_count
96
96
phi3.rope.freq_base
10000
10000
phi3.rope.scaling.attn_factor
1.1902381
1.1902381
phi3.rope.scaling.original_context_length
4096
4096
tokenizer.ggml.add_bos_token
false
false
tokenizer.ggml.add_eos_token
false
false
tokenizer.ggml.bos_token_id
1
1
tokenizer.ggml.eos_token_id
32000
32000
tokenizer.ggml.model
llama
llama
tokenizer.ggml.padding_token_id
32000
32000
tokenizer.ggml.pre
default
default
tokenizer.ggml.scores
[-1000, -1000, -1000, 0, 0, ...]
[-1000, -1000, -1000, 0, 0, ...]
tokenizer.ggml.token_type
[3, 3, 4, 6, 6, ...]
[3, 3, 4, 6, 6, ...]
tokenizer.ggml.tokens
[<unk>, <s>, </s>, <0x00>, <0x01>, ...]
[<unk>, <s>, </s>, <0x00>, <0x01>, ...]
tokenizer.ggml.unknown_token_id
0
0
Tensor
Name
Type
Shape
token_embd.weight
Q4_0
Q4_0
[3072, 32064]
blk.0
blk.0.attn_norm.weight
F32
F32
[3072]
blk.0.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.0.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.0.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.0.ffn_norm.weight
F32
F32
[3072]
blk.0.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.1
blk.1.attn_norm.weight
F32
F32
[3072]
blk.1.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.1.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.1.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.1.ffn_norm.weight
F32
F32
[3072]
blk.1.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.2
blk.2.attn_norm.weight
F32
F32
[3072]
blk.2.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.2.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.2.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.2.ffn_norm.weight
F32
F32
[3072]
blk.2.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.3
blk.3.attn_norm.weight
F32
F32
[3072]
blk.3.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.3.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.3.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.3.ffn_norm.weight
F32
F32
[3072]
blk.3.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.4
blk.4.attn_norm.weight
F32
F32
[3072]
blk.4.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.4.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.4.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.4.ffn_norm.weight
F32
F32
[3072]
blk.4.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.5
blk.5.attn_norm.weight
F32
F32
[3072]
blk.5.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.5.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.5.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.5.ffn_norm.weight
F32
F32
[3072]
blk.5.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.6
blk.6.attn_norm.weight
F32
F32
[3072]
blk.6.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.6.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.6.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.6.ffn_norm.weight
F32
F32
[3072]
blk.6.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.7
blk.7.attn_norm.weight
F32
F32
[3072]
blk.7.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.7.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.7.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.7.ffn_norm.weight
F32
F32
[3072]
blk.7.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.8
blk.8.attn_norm.weight
F32
F32
[3072]
blk.8.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.8.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.8.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.8.ffn_norm.weight
F32
F32
[3072]
blk.8.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.9
blk.9.attn_norm.weight
F32
F32
[3072]
blk.9.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.9.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.9.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.9.ffn_norm.weight
F32
F32
[3072]
blk.9.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.10
blk.10.attn_norm.weight
F32
F32
[3072]
blk.10.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.10.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.10.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.10.ffn_norm.weight
F32
F32
[3072]
blk.10.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.11
blk.11.attn_norm.weight
F32
F32
[3072]
blk.11.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.11.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.11.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.11.ffn_norm.weight
F32
F32
[3072]
blk.11.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.12
blk.12.attn_norm.weight
F32
F32
[3072]
blk.12.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.12.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.12.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.12.ffn_norm.weight
F32
F32
[3072]
blk.12.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.13
blk.13.attn_norm.weight
F32
F32
[3072]
blk.13.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.13.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.13.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.13.ffn_norm.weight
F32
F32
[3072]
blk.13.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.14
blk.14.attn_norm.weight
F32
F32
[3072]
blk.14.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.14.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.14.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.14.ffn_norm.weight
F32
F32
[3072]
blk.14.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.15
blk.15.attn_norm.weight
F32
F32
[3072]
blk.15.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.15.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.15.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.15.ffn_norm.weight
F32
F32
[3072]
blk.15.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.16
blk.16.attn_norm.weight
F32
F32
[3072]
blk.16.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.16.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.16.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.16.ffn_norm.weight
F32
F32
[3072]
blk.16.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.17
blk.17.attn_norm.weight
F32
F32
[3072]
blk.17.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.17.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.17.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.17.ffn_norm.weight
F32
F32
[3072]
blk.17.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.18
blk.18.attn_norm.weight
F32
F32
[3072]
blk.18.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.18.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.18.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.18.ffn_norm.weight
F32
F32
[3072]
blk.18.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.19
blk.19.attn_norm.weight
F32
F32
[3072]
blk.19.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.19.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.19.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.19.ffn_norm.weight
F32
F32
[3072]
blk.19.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.20
blk.20.attn_norm.weight
F32
F32
[3072]
blk.20.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.20.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.20.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.20.ffn_norm.weight
F32
F32
[3072]
blk.20.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.21
blk.21.attn_norm.weight
F32
F32
[3072]
blk.21.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.21.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.21.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.21.ffn_norm.weight
F32
F32
[3072]
blk.21.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.22
blk.22.attn_norm.weight
F32
F32
[3072]
blk.22.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.22.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.22.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.22.ffn_norm.weight
F32
F32
[3072]
blk.22.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.23
blk.23.attn_norm.weight
F32
F32
[3072]
blk.23.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.23.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.23.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.23.ffn_norm.weight
F32
F32
[3072]
blk.23.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.24
blk.24.attn_norm.weight
F32
F32
[3072]
blk.24.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.24.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.24.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.24.ffn_norm.weight
F32
F32
[3072]
blk.24.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.25
blk.25.attn_norm.weight
F32
F32
[3072]
blk.25.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.25.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.25.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.25.ffn_norm.weight
F32
F32
[3072]
blk.25.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.26
blk.26.attn_norm.weight
F32
F32
[3072]
blk.26.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.26.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.26.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.26.ffn_norm.weight
F32
F32
[3072]
blk.26.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.27
blk.27.attn_norm.weight
F32
F32
[3072]
blk.27.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.27.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.27.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.27.ffn_norm.weight
F32
F32
[3072]
blk.27.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.28
blk.28.attn_norm.weight
F32
F32
[3072]
blk.28.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.28.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.28.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.28.ffn_norm.weight
F32
F32
[3072]
blk.28.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.29
blk.29.attn_norm.weight
F32
F32
[3072]
blk.29.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.29.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.29.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.29.ffn_norm.weight
F32
F32
[3072]
blk.29.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.30
blk.30.attn_norm.weight
F32
F32
[3072]
blk.30.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.30.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.30.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.30.ffn_norm.weight
F32
F32
[3072]
blk.30.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
blk.31
blk.31.attn_norm.weight
F32
F32
[3072]
blk.31.attn_output.weight
Q4_0
Q4_0
[3072, 3072]
blk.31.attn_qkv.weight
Q4_0
Q4_0
[3072, 9216]
blk.31.ffn_down.weight
Q4_0
Q4_0
[8192, 3072]
blk.31.ffn_norm.weight
F32
F32
[3072]
blk.31.ffn_up.weight
Q4_0
Q4_0
[3072, 16384]
output.weight
Q6_K
Q6_K
[3072, 32064]
rope_factors_long.weight
F32
F32
[48]
rope_factors_short.weight
F32
F32
[48]
output_norm.weight
F32
F32
[3072]