Models
GitHub
Discord
Turbo
Sign in
Download
Models
Download
GitHub
Discord
Sign in
opencoder
:1.5b-instruct-fp16
129.1K
Downloads
Updated
9 months ago
OpenCoder is an open and reproducible code LLM family which includes 1.5B and 8B models, supporting chat in English and Chinese languages.
OpenCoder is an open and reproducible code LLM family which includes 1.5B and 8B models, supporting chat in English and Chinese languages.
Cancel
1.5b
8b
opencoder:1.5b-instruct-fp16
...
/
model
8fa7b1ae0d10 · 3.8GB
Metadata
general.architecture
llama
llama
general.file_type
F16
F16
llama.attention.head_count
14
14
llama.attention.head_count_kv
14
14
llama.attention.layer_norm_rms_epsilon
1e-05
1e-05
llama.block_count
24
24
llama.context_length
4096
4096
llama.embedding_length
2240
2240
llama.feed_forward_length
6144
6144
llama.rope.dimension_count
160
160
llama.rope.freq_base
10000
10000
llama.vocab_size
96640
96640
tokenizer.ggml.add_bos_token
false
false
tokenizer.ggml.add_eos_token
false
false
tokenizer.ggml.add_space_prefix
false
false
tokenizer.ggml.bos_token_id
96540
96540
tokenizer.ggml.eos_token_id
96539
96539
tokenizer.ggml.model
llama
llama
tokenizer.ggml.pre
default
default
tokenizer.ggml.scores
[-1000, -1000, -1000, -1000, 0, ...]
[-1000, -1000, -1000, -1000, 0, ...]
tokenizer.ggml.token_type
[3, 3, 3, 3, 6, ...]
[3, 3, 3, 3, 6, ...]
tokenizer.ggml.tokens
[<unk>, <s>, </s>, <pad>, <0x00>, ...]
[<unk>, <s>, </s>, <pad>, <0x00>, ...]
Tensor
Name
Type
Shape
token_embd.weight
F16
F16
[2240, 96640]
blk.0
blk.0.attn_k.weight
F16
F16
[2240, 2240]
blk.0.attn_norm.weight
F32
F32
[2240]
blk.0.attn_output.weight
F16
F16
[2240, 2240]
blk.0.attn_q.weight
F16
F16
[2240, 2240]
blk.0.attn_v.weight
F16
F16
[2240, 2240]
blk.0.ffn_down.weight
F16
F16
[6144, 2240]
blk.0.ffn_gate.weight
F16
F16
[2240, 6144]
blk.0.ffn_norm.weight
F32
F32
[2240]
blk.0.ffn_up.weight
F16
F16
[2240, 6144]
blk.1
blk.1.attn_k.weight
F16
F16
[2240, 2240]
blk.1.attn_norm.weight
F32
F32
[2240]
blk.1.attn_output.weight
F16
F16
[2240, 2240]
blk.1.attn_q.weight
F16
F16
[2240, 2240]
blk.1.attn_v.weight
F16
F16
[2240, 2240]
blk.1.ffn_down.weight
F16
F16
[6144, 2240]
blk.1.ffn_gate.weight
F16
F16
[2240, 6144]
blk.1.ffn_norm.weight
F32
F32
[2240]
blk.1.ffn_up.weight
F16
F16
[2240, 6144]
blk.2
blk.2.attn_k.weight
F16
F16
[2240, 2240]
blk.2.attn_norm.weight
F32
F32
[2240]
blk.2.attn_output.weight
F16
F16
[2240, 2240]
blk.2.attn_q.weight
F16
F16
[2240, 2240]
blk.2.attn_v.weight
F16
F16
[2240, 2240]
blk.2.ffn_down.weight
F16
F16
[6144, 2240]
blk.2.ffn_gate.weight
F16
F16
[2240, 6144]
blk.2.ffn_norm.weight
F32
F32
[2240]
blk.2.ffn_up.weight
F16
F16
[2240, 6144]
blk.3
blk.3.attn_k.weight
F16
F16
[2240, 2240]
blk.3.attn_norm.weight
F32
F32
[2240]
blk.3.attn_output.weight
F16
F16
[2240, 2240]
blk.3.attn_q.weight
F16
F16
[2240, 2240]
blk.3.attn_v.weight
F16
F16
[2240, 2240]
blk.3.ffn_down.weight
F16
F16
[6144, 2240]
blk.3.ffn_gate.weight
F16
F16
[2240, 6144]
blk.3.ffn_norm.weight
F32
F32
[2240]
blk.3.ffn_up.weight
F16
F16
[2240, 6144]
blk.4
blk.4.attn_k.weight
F16
F16
[2240, 2240]
blk.4.attn_norm.weight
F32
F32
[2240]
blk.4.attn_output.weight
F16
F16
[2240, 2240]
blk.4.attn_q.weight
F16
F16
[2240, 2240]
blk.4.attn_v.weight
F16
F16
[2240, 2240]
blk.4.ffn_down.weight
F16
F16
[6144, 2240]
blk.4.ffn_gate.weight
F16
F16
[2240, 6144]
blk.4.ffn_norm.weight
F32
F32
[2240]
blk.4.ffn_up.weight
F16
F16
[2240, 6144]
blk.5
blk.5.attn_k.weight
F16
F16
[2240, 2240]
blk.5.attn_norm.weight
F32
F32
[2240]
blk.5.attn_output.weight
F16
F16
[2240, 2240]
blk.5.attn_q.weight
F16
F16
[2240, 2240]
blk.5.attn_v.weight
F16
F16
[2240, 2240]
blk.5.ffn_down.weight
F16
F16
[6144, 2240]
blk.5.ffn_gate.weight
F16
F16
[2240, 6144]
blk.5.ffn_norm.weight
F32
F32
[2240]
blk.5.ffn_up.weight
F16
F16
[2240, 6144]
blk.6
blk.6.attn_k.weight
F16
F16
[2240, 2240]
blk.6.attn_norm.weight
F32
F32
[2240]
blk.6.attn_output.weight
F16
F16
[2240, 2240]
blk.6.attn_q.weight
F16
F16
[2240, 2240]
blk.6.attn_v.weight
F16
F16
[2240, 2240]
blk.6.ffn_down.weight
F16
F16
[6144, 2240]
blk.6.ffn_gate.weight
F16
F16
[2240, 6144]
blk.6.ffn_norm.weight
F32
F32
[2240]
blk.6.ffn_up.weight
F16
F16
[2240, 6144]
blk.7
blk.7.attn_k.weight
F16
F16
[2240, 2240]
blk.7.attn_norm.weight
F32
F32
[2240]
blk.7.attn_output.weight
F16
F16
[2240, 2240]
blk.7.attn_q.weight
F16
F16
[2240, 2240]
blk.7.attn_v.weight
F16
F16
[2240, 2240]
blk.7.ffn_down.weight
F16
F16
[6144, 2240]
blk.7.ffn_gate.weight
F16
F16
[2240, 6144]
blk.7.ffn_norm.weight
F32
F32
[2240]
blk.7.ffn_up.weight
F16
F16
[2240, 6144]
blk.8
blk.8.attn_k.weight
F16
F16
[2240, 2240]
blk.8.attn_norm.weight
F32
F32
[2240]
blk.8.attn_output.weight
F16
F16
[2240, 2240]
blk.8.attn_q.weight
F16
F16
[2240, 2240]
blk.8.attn_v.weight
F16
F16
[2240, 2240]
blk.8.ffn_down.weight
F16
F16
[6144, 2240]
blk.8.ffn_gate.weight
F16
F16
[2240, 6144]
blk.8.ffn_norm.weight
F32
F32
[2240]
blk.8.ffn_up.weight
F16
F16
[2240, 6144]
blk.9
blk.9.attn_k.weight
F16
F16
[2240, 2240]
blk.9.attn_norm.weight
F32
F32
[2240]
blk.9.attn_output.weight
F16
F16
[2240, 2240]
blk.9.attn_q.weight
F16
F16
[2240, 2240]
blk.9.attn_v.weight
F16
F16
[2240, 2240]
blk.9.ffn_down.weight
F16
F16
[6144, 2240]
blk.9.ffn_gate.weight
F16
F16
[2240, 6144]
blk.9.ffn_norm.weight
F32
F32
[2240]
blk.9.ffn_up.weight
F16
F16
[2240, 6144]
blk.10
blk.10.attn_k.weight
F16
F16
[2240, 2240]
blk.10.attn_norm.weight
F32
F32
[2240]
blk.10.attn_output.weight
F16
F16
[2240, 2240]
blk.10.attn_q.weight
F16
F16
[2240, 2240]
blk.10.attn_v.weight
F16
F16
[2240, 2240]
blk.10.ffn_down.weight
F16
F16
[6144, 2240]
blk.10.ffn_gate.weight
F16
F16
[2240, 6144]
blk.10.ffn_norm.weight
F32
F32
[2240]
blk.10.ffn_up.weight
F16
F16
[2240, 6144]
blk.11
blk.11.attn_k.weight
F16
F16
[2240, 2240]
blk.11.attn_norm.weight
F32
F32
[2240]
blk.11.attn_output.weight
F16
F16
[2240, 2240]
blk.11.attn_q.weight
F16
F16
[2240, 2240]
blk.11.attn_v.weight
F16
F16
[2240, 2240]
blk.11.ffn_down.weight
F16
F16
[6144, 2240]
blk.11.ffn_gate.weight
F16
F16
[2240, 6144]
blk.11.ffn_norm.weight
F32
F32
[2240]
blk.11.ffn_up.weight
F16
F16
[2240, 6144]
blk.12
blk.12.attn_k.weight
F16
F16
[2240, 2240]
blk.12.attn_norm.weight
F32
F32
[2240]
blk.12.attn_output.weight
F16
F16
[2240, 2240]
blk.12.attn_q.weight
F16
F16
[2240, 2240]
blk.12.attn_v.weight
F16
F16
[2240, 2240]
blk.12.ffn_down.weight
F16
F16
[6144, 2240]
blk.12.ffn_gate.weight
F16
F16
[2240, 6144]
blk.12.ffn_norm.weight
F32
F32
[2240]
blk.12.ffn_up.weight
F16
F16
[2240, 6144]
blk.13
blk.13.attn_k.weight
F16
F16
[2240, 2240]
blk.13.attn_norm.weight
F32
F32
[2240]
blk.13.attn_output.weight
F16
F16
[2240, 2240]
blk.13.attn_q.weight
F16
F16
[2240, 2240]
blk.13.attn_v.weight
F16
F16
[2240, 2240]
blk.13.ffn_down.weight
F16
F16
[6144, 2240]
blk.13.ffn_gate.weight
F16
F16
[2240, 6144]
blk.13.ffn_norm.weight
F32
F32
[2240]
blk.13.ffn_up.weight
F16
F16
[2240, 6144]
blk.14
blk.14.attn_k.weight
F16
F16
[2240, 2240]
blk.14.attn_norm.weight
F32
F32
[2240]
blk.14.attn_output.weight
F16
F16
[2240, 2240]
blk.14.attn_q.weight
F16
F16
[2240, 2240]
blk.14.attn_v.weight
F16
F16
[2240, 2240]
blk.14.ffn_down.weight
F16
F16
[6144, 2240]
blk.14.ffn_gate.weight
F16
F16
[2240, 6144]
blk.14.ffn_norm.weight
F32
F32
[2240]
blk.14.ffn_up.weight
F16
F16
[2240, 6144]
blk.15
blk.15.attn_k.weight
F16
F16
[2240, 2240]
blk.15.attn_norm.weight
F32
F32
[2240]
blk.15.attn_output.weight
F16
F16
[2240, 2240]
blk.15.attn_q.weight
F16
F16
[2240, 2240]
blk.15.attn_v.weight
F16
F16
[2240, 2240]
blk.15.ffn_down.weight
F16
F16
[6144, 2240]
blk.15.ffn_gate.weight
F16
F16
[2240, 6144]
blk.15.ffn_norm.weight
F32
F32
[2240]
blk.15.ffn_up.weight
F16
F16
[2240, 6144]
blk.16
blk.16.attn_k.weight
F16
F16
[2240, 2240]
blk.16.attn_norm.weight
F32
F32
[2240]
blk.16.attn_output.weight
F16
F16
[2240, 2240]
blk.16.attn_q.weight
F16
F16
[2240, 2240]
blk.16.attn_v.weight
F16
F16
[2240, 2240]
blk.16.ffn_down.weight
F16
F16
[6144, 2240]
blk.16.ffn_gate.weight
F16
F16
[2240, 6144]
blk.16.ffn_norm.weight
F32
F32
[2240]
blk.16.ffn_up.weight
F16
F16
[2240, 6144]
blk.17
blk.17.attn_k.weight
F16
F16
[2240, 2240]
blk.17.attn_norm.weight
F32
F32
[2240]
blk.17.attn_output.weight
F16
F16
[2240, 2240]
blk.17.attn_q.weight
F16
F16
[2240, 2240]
blk.17.attn_v.weight
F16
F16
[2240, 2240]
blk.17.ffn_down.weight
F16
F16
[6144, 2240]
blk.17.ffn_gate.weight
F16
F16
[2240, 6144]
blk.17.ffn_norm.weight
F32
F32
[2240]
blk.17.ffn_up.weight
F16
F16
[2240, 6144]
blk.18
blk.18.attn_k.weight
F16
F16
[2240, 2240]
blk.18.attn_norm.weight
F32
F32
[2240]
blk.18.attn_output.weight
F16
F16
[2240, 2240]
blk.18.attn_q.weight
F16
F16
[2240, 2240]
blk.18.attn_v.weight
F16
F16
[2240, 2240]
blk.18.ffn_down.weight
F16
F16
[6144, 2240]
blk.18.ffn_gate.weight
F16
F16
[2240, 6144]
blk.18.ffn_norm.weight
F32
F32
[2240]
blk.18.ffn_up.weight
F16
F16
[2240, 6144]
blk.19
blk.19.attn_k.weight
F16
F16
[2240, 2240]
blk.19.attn_norm.weight
F32
F32
[2240]
blk.19.attn_output.weight
F16
F16
[2240, 2240]
blk.19.attn_q.weight
F16
F16
[2240, 2240]
blk.19.attn_v.weight
F16
F16
[2240, 2240]
blk.19.ffn_down.weight
F16
F16
[6144, 2240]
blk.19.ffn_gate.weight
F16
F16
[2240, 6144]
blk.19.ffn_norm.weight
F32
F32
[2240]
blk.19.ffn_up.weight
F16
F16
[2240, 6144]
blk.20
blk.20.attn_k.weight
F16
F16
[2240, 2240]
blk.20.attn_norm.weight
F32
F32
[2240]
blk.20.attn_output.weight
F16
F16
[2240, 2240]
blk.20.attn_q.weight
F16
F16
[2240, 2240]
blk.20.attn_v.weight
F16
F16
[2240, 2240]
blk.20.ffn_down.weight
F16
F16
[6144, 2240]
blk.20.ffn_gate.weight
F16
F16
[2240, 6144]
blk.20.ffn_norm.weight
F32
F32
[2240]
blk.20.ffn_up.weight
F16
F16
[2240, 6144]
blk.21
blk.21.attn_k.weight
F16
F16
[2240, 2240]
blk.21.attn_norm.weight
F32
F32
[2240]
blk.21.attn_output.weight
F16
F16
[2240, 2240]
blk.21.attn_q.weight
F16
F16
[2240, 2240]
blk.21.attn_v.weight
F16
F16
[2240, 2240]
blk.21.ffn_down.weight
F16
F16
[6144, 2240]
blk.21.ffn_gate.weight
F16
F16
[2240, 6144]
blk.21.ffn_norm.weight
F32
F32
[2240]
blk.21.ffn_up.weight
F16
F16
[2240, 6144]
blk.22
blk.22.attn_k.weight
F16
F16
[2240, 2240]
blk.22.attn_norm.weight
F32
F32
[2240]
blk.22.attn_output.weight
F16
F16
[2240, 2240]
blk.22.attn_q.weight
F16
F16
[2240, 2240]
blk.22.attn_v.weight
F16
F16
[2240, 2240]
blk.22.ffn_down.weight
F16
F16
[6144, 2240]
blk.22.ffn_gate.weight
F16
F16
[2240, 6144]
blk.22.ffn_norm.weight
F32
F32
[2240]
blk.22.ffn_up.weight
F16
F16
[2240, 6144]
blk.23
blk.23.attn_k.weight
F16
F16
[2240, 2240]
blk.23.attn_norm.weight
F32
F32
[2240]
blk.23.attn_output.weight
F16
F16
[2240, 2240]
blk.23.attn_q.weight
F16
F16
[2240, 2240]
blk.23.attn_v.weight
F16
F16
[2240, 2240]
blk.23.ffn_down.weight
F16
F16
[6144, 2240]
blk.23.ffn_gate.weight
F16
F16
[2240, 6144]
blk.23.ffn_norm.weight
F32
F32
[2240]
blk.23.ffn_up.weight
F16
F16
[2240, 6144]
output.weight
F16
F16
[2240, 96640]
output_norm.weight
F32
F32
[2240]