Models
GitHub
Discord
Turbo
Sign in
Download
Models
Download
GitHub
Discord
Sign in
dolphin-phi
:2.7b-v2.6-q5_0
246.5K
Downloads
Updated
1 year ago
2.7B uncensored Dolphin model by Eric Hartford, based on the Phi language model by Microsoft Research.
2.7B uncensored Dolphin model by Eric Hartford, based on the Phi language model by Microsoft Research.
Cancel
2.7b
dolphin-phi:2.7b-v2.6-q5_0
...
/
model
4a17736fbf8c · 1.9GB
Metadata
general.architecture
phi2
phi2
general.file_type
Q5_0
Q5_0
phi2.attention.head_count
32
32
phi2.attention.head_count_kv
32
32
phi2.attention.layer_norm_epsilon
1e-05
1e-05
phi2.block_count
32
32
phi2.context_length
2048
2048
phi2.embedding_length
2560
2560
phi2.feed_forward_length
10240
10240
phi2.rope.dimension_count
32
32
tokenizer.ggml.add_bos_token
false
false
tokenizer.ggml.bos_token_id
50256
50256
tokenizer.ggml.eos_token_id
50295
50295
tokenizer.ggml.merges
[Ġ t, Ġ a, h e, i n, r e, ...]
[Ġ t, Ġ a, h e, i n, r e, ...]
tokenizer.ggml.model
gpt2
gpt2
tokenizer.ggml.padding_token_id
50256
50256
tokenizer.ggml.token_type
[1, 1, 1, 1, 1, ...]
[1, 1, 1, 1, 1, ...]
tokenizer.ggml.tokens
[!, ", #, $, %, ...]
[!, ", #, $, %, ...]
tokenizer.ggml.unknown_token_id
50256
50256
Tensor
Name
Type
Shape
token_embd.weight
Q5_0
Q5_0
[2560, 51200]
blk.0
blk.0.attn_norm.bias
F32
F32
[2560]
blk.0.attn_norm.weight
F32
F32
[2560]
blk.0.attn_output.bias
F32
F32
[2560]
blk.0.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.0.attn_qkv.bias
F32
F32
[7680]
blk.0.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.0.ffn_down.bias
F32
F32
[2560]
blk.0.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.0.ffn_up.bias
F32
F32
[10240]
blk.0.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.1
blk.1.attn_norm.bias
F32
F32
[2560]
blk.1.attn_norm.weight
F32
F32
[2560]
blk.1.attn_output.bias
F32
F32
[2560]
blk.1.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.1.attn_qkv.bias
F32
F32
[7680]
blk.1.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.1.ffn_down.bias
F32
F32
[2560]
blk.1.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.1.ffn_up.bias
F32
F32
[10240]
blk.1.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.2
blk.2.attn_norm.bias
F32
F32
[2560]
blk.2.attn_norm.weight
F32
F32
[2560]
blk.2.attn_output.bias
F32
F32
[2560]
blk.2.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.2.attn_qkv.bias
F32
F32
[7680]
blk.2.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.2.ffn_down.bias
F32
F32
[2560]
blk.2.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.2.ffn_up.bias
F32
F32
[10240]
blk.2.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.3
blk.3.attn_norm.bias
F32
F32
[2560]
blk.3.attn_norm.weight
F32
F32
[2560]
blk.3.attn_output.bias
F32
F32
[2560]
blk.3.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.3.attn_qkv.bias
F32
F32
[7680]
blk.3.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.3.ffn_down.bias
F32
F32
[2560]
blk.3.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.3.ffn_up.bias
F32
F32
[10240]
blk.3.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.4
blk.4.attn_norm.bias
F32
F32
[2560]
blk.4.attn_norm.weight
F32
F32
[2560]
blk.4.attn_output.bias
F32
F32
[2560]
blk.4.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.4.attn_qkv.bias
F32
F32
[7680]
blk.4.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.4.ffn_down.bias
F32
F32
[2560]
blk.4.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.4.ffn_up.bias
F32
F32
[10240]
blk.4.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.5
blk.5.attn_norm.bias
F32
F32
[2560]
blk.5.attn_norm.weight
F32
F32
[2560]
blk.5.attn_output.bias
F32
F32
[2560]
blk.5.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.5.attn_qkv.bias
F32
F32
[7680]
blk.5.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.5.ffn_down.bias
F32
F32
[2560]
blk.5.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.5.ffn_up.bias
F32
F32
[10240]
blk.5.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.6
blk.6.attn_norm.bias
F32
F32
[2560]
blk.6.attn_norm.weight
F32
F32
[2560]
blk.6.attn_output.bias
F32
F32
[2560]
blk.6.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.6.attn_qkv.bias
F32
F32
[7680]
blk.6.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.6.ffn_down.bias
F32
F32
[2560]
blk.6.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.6.ffn_up.bias
F32
F32
[10240]
blk.6.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.7
blk.7.attn_norm.bias
F32
F32
[2560]
blk.7.attn_norm.weight
F32
F32
[2560]
blk.7.attn_output.bias
F32
F32
[2560]
blk.7.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.7.attn_qkv.bias
F32
F32
[7680]
blk.7.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.7.ffn_down.bias
F32
F32
[2560]
blk.7.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.7.ffn_up.bias
F32
F32
[10240]
blk.7.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.8
blk.8.attn_norm.bias
F32
F32
[2560]
blk.8.attn_norm.weight
F32
F32
[2560]
blk.8.attn_output.bias
F32
F32
[2560]
blk.8.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.8.attn_qkv.bias
F32
F32
[7680]
blk.8.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.8.ffn_down.bias
F32
F32
[2560]
blk.8.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.8.ffn_up.bias
F32
F32
[10240]
blk.8.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.9
blk.9.attn_norm.bias
F32
F32
[2560]
blk.9.attn_norm.weight
F32
F32
[2560]
blk.9.attn_output.bias
F32
F32
[2560]
blk.9.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.9.attn_qkv.bias
F32
F32
[7680]
blk.9.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.9.ffn_down.bias
F32
F32
[2560]
blk.9.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.9.ffn_up.bias
F32
F32
[10240]
blk.9.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.10
blk.10.attn_norm.bias
F32
F32
[2560]
blk.10.attn_norm.weight
F32
F32
[2560]
blk.10.attn_output.bias
F32
F32
[2560]
blk.10.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.10.attn_qkv.bias
F32
F32
[7680]
blk.10.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.10.ffn_down.bias
F32
F32
[2560]
blk.10.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.10.ffn_up.bias
F32
F32
[10240]
blk.10.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.11
blk.11.attn_norm.bias
F32
F32
[2560]
blk.11.attn_norm.weight
F32
F32
[2560]
blk.11.attn_output.bias
F32
F32
[2560]
blk.11.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.11.attn_qkv.bias
F32
F32
[7680]
blk.11.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.11.ffn_down.bias
F32
F32
[2560]
blk.11.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.11.ffn_up.bias
F32
F32
[10240]
blk.11.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.12
blk.12.attn_norm.bias
F32
F32
[2560]
blk.12.attn_norm.weight
F32
F32
[2560]
blk.12.attn_output.bias
F32
F32
[2560]
blk.12.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.12.attn_qkv.bias
F32
F32
[7680]
blk.12.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.12.ffn_down.bias
F32
F32
[2560]
blk.12.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.12.ffn_up.bias
F32
F32
[10240]
blk.12.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.13
blk.13.attn_norm.bias
F32
F32
[2560]
blk.13.attn_norm.weight
F32
F32
[2560]
blk.13.attn_output.bias
F32
F32
[2560]
blk.13.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.13.attn_qkv.bias
F32
F32
[7680]
blk.13.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.13.ffn_down.bias
F32
F32
[2560]
blk.13.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.13.ffn_up.bias
F32
F32
[10240]
blk.13.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.14
blk.14.attn_norm.bias
F32
F32
[2560]
blk.14.attn_norm.weight
F32
F32
[2560]
blk.14.attn_output.bias
F32
F32
[2560]
blk.14.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.14.attn_qkv.bias
F32
F32
[7680]
blk.14.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.14.ffn_down.bias
F32
F32
[2560]
blk.14.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.14.ffn_up.bias
F32
F32
[10240]
blk.14.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.15
blk.15.attn_norm.bias
F32
F32
[2560]
blk.15.attn_norm.weight
F32
F32
[2560]
blk.15.attn_output.bias
F32
F32
[2560]
blk.15.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.15.attn_qkv.bias
F32
F32
[7680]
blk.15.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.15.ffn_down.bias
F32
F32
[2560]
blk.15.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.15.ffn_up.bias
F32
F32
[10240]
blk.15.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.16
blk.16.attn_norm.bias
F32
F32
[2560]
blk.16.attn_norm.weight
F32
F32
[2560]
blk.16.attn_output.bias
F32
F32
[2560]
blk.16.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.16.attn_qkv.bias
F32
F32
[7680]
blk.16.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.16.ffn_down.bias
F32
F32
[2560]
blk.16.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.16.ffn_up.bias
F32
F32
[10240]
blk.16.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.17
blk.17.attn_norm.bias
F32
F32
[2560]
blk.17.attn_norm.weight
F32
F32
[2560]
blk.17.attn_output.bias
F32
F32
[2560]
blk.17.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.17.attn_qkv.bias
F32
F32
[7680]
blk.17.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.17.ffn_down.bias
F32
F32
[2560]
blk.17.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.17.ffn_up.bias
F32
F32
[10240]
blk.17.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.18
blk.18.attn_norm.bias
F32
F32
[2560]
blk.18.attn_norm.weight
F32
F32
[2560]
blk.18.attn_output.bias
F32
F32
[2560]
blk.18.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.18.attn_qkv.bias
F32
F32
[7680]
blk.18.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.18.ffn_down.bias
F32
F32
[2560]
blk.18.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.18.ffn_up.bias
F32
F32
[10240]
blk.18.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.19
blk.19.attn_norm.bias
F32
F32
[2560]
blk.19.attn_norm.weight
F32
F32
[2560]
blk.19.attn_output.bias
F32
F32
[2560]
blk.19.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.19.attn_qkv.bias
F32
F32
[7680]
blk.19.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.19.ffn_down.bias
F32
F32
[2560]
blk.19.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.19.ffn_up.bias
F32
F32
[10240]
blk.19.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.20
blk.20.attn_norm.bias
F32
F32
[2560]
blk.20.attn_norm.weight
F32
F32
[2560]
blk.20.attn_output.bias
F32
F32
[2560]
blk.20.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.20.attn_qkv.bias
F32
F32
[7680]
blk.20.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.20.ffn_down.bias
F32
F32
[2560]
blk.20.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.20.ffn_up.bias
F32
F32
[10240]
blk.20.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.21
blk.21.attn_norm.bias
F32
F32
[2560]
blk.21.attn_norm.weight
F32
F32
[2560]
blk.21.attn_output.bias
F32
F32
[2560]
blk.21.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.21.attn_qkv.bias
F32
F32
[7680]
blk.21.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.21.ffn_down.bias
F32
F32
[2560]
blk.21.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.21.ffn_up.bias
F32
F32
[10240]
blk.21.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.22
blk.22.attn_norm.bias
F32
F32
[2560]
blk.22.attn_norm.weight
F32
F32
[2560]
blk.22.attn_output.bias
F32
F32
[2560]
blk.22.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.22.attn_qkv.bias
F32
F32
[7680]
blk.22.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.22.ffn_down.bias
F32
F32
[2560]
blk.22.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.22.ffn_up.bias
F32
F32
[10240]
blk.22.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.23
blk.23.attn_norm.bias
F32
F32
[2560]
blk.23.attn_norm.weight
F32
F32
[2560]
blk.23.attn_output.bias
F32
F32
[2560]
blk.23.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.23.attn_qkv.bias
F32
F32
[7680]
blk.23.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.23.ffn_down.bias
F32
F32
[2560]
blk.23.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.23.ffn_up.bias
F32
F32
[10240]
blk.23.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.24
blk.24.attn_norm.bias
F32
F32
[2560]
blk.24.attn_norm.weight
F32
F32
[2560]
blk.24.attn_output.bias
F32
F32
[2560]
blk.24.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.24.attn_qkv.bias
F32
F32
[7680]
blk.24.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.24.ffn_down.bias
F32
F32
[2560]
blk.24.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.24.ffn_up.bias
F32
F32
[10240]
blk.24.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.25
blk.25.attn_norm.bias
F32
F32
[2560]
blk.25.attn_norm.weight
F32
F32
[2560]
blk.25.attn_output.bias
F32
F32
[2560]
blk.25.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.25.attn_qkv.bias
F32
F32
[7680]
blk.25.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.25.ffn_down.bias
F32
F32
[2560]
blk.25.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.25.ffn_up.bias
F32
F32
[10240]
blk.25.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.26
blk.26.attn_norm.bias
F32
F32
[2560]
blk.26.attn_norm.weight
F32
F32
[2560]
blk.26.attn_output.bias
F32
F32
[2560]
blk.26.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.26.attn_qkv.bias
F32
F32
[7680]
blk.26.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.26.ffn_down.bias
F32
F32
[2560]
blk.26.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.26.ffn_up.bias
F32
F32
[10240]
blk.26.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.27
blk.27.attn_norm.bias
F32
F32
[2560]
blk.27.attn_norm.weight
F32
F32
[2560]
blk.27.attn_output.bias
F32
F32
[2560]
blk.27.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.27.attn_qkv.bias
F32
F32
[7680]
blk.27.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.27.ffn_down.bias
F32
F32
[2560]
blk.27.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.27.ffn_up.bias
F32
F32
[10240]
blk.27.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.28
blk.28.attn_norm.bias
F32
F32
[2560]
blk.28.attn_norm.weight
F32
F32
[2560]
blk.28.attn_output.bias
F32
F32
[2560]
blk.28.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.28.attn_qkv.bias
F32
F32
[7680]
blk.28.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.28.ffn_down.bias
F32
F32
[2560]
blk.28.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.28.ffn_up.bias
F32
F32
[10240]
blk.28.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.29
blk.29.attn_norm.bias
F32
F32
[2560]
blk.29.attn_norm.weight
F32
F32
[2560]
blk.29.attn_output.bias
F32
F32
[2560]
blk.29.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.29.attn_qkv.bias
F32
F32
[7680]
blk.29.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.29.ffn_down.bias
F32
F32
[2560]
blk.29.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.29.ffn_up.bias
F32
F32
[10240]
blk.29.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.30
blk.30.attn_norm.bias
F32
F32
[2560]
blk.30.attn_norm.weight
F32
F32
[2560]
blk.30.attn_output.bias
F32
F32
[2560]
blk.30.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.30.attn_qkv.bias
F32
F32
[7680]
blk.30.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.30.ffn_down.bias
F32
F32
[2560]
blk.30.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.30.ffn_up.bias
F32
F32
[10240]
blk.30.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
blk.31
blk.31.attn_norm.bias
F32
F32
[2560]
blk.31.attn_norm.weight
F32
F32
[2560]
blk.31.attn_output.bias
F32
F32
[2560]
blk.31.attn_output.weight
Q5_0
Q5_0
[2560, 2560]
blk.31.attn_qkv.bias
F32
F32
[7680]
blk.31.attn_qkv.weight
Q5_0
Q5_0
[2560, 7680]
blk.31.ffn_down.bias
F32
F32
[2560]
blk.31.ffn_down.weight
Q5_0
Q5_0
[10240, 2560]
blk.31.ffn_up.bias
F32
F32
[10240]
blk.31.ffn_up.weight
Q5_0
Q5_0
[2560, 10240]
output.bias
F32
F32
[51200]
output.weight
Q6_K
Q6_K
[2560, 51200]
output_norm.bias
F32
F32
[2560]
output_norm.weight
F32
F32
[2560]