StarCoder2 is the next generation of transparently trained open code LLMs that comes in three sizes: 3B, 7B and 15B parameters.
3b
7b
15b
438.5K Pulls Updated 2 months ago
Updated 2 months ago
2 months ago
9f4ae0aff61e · 1.7GB
model
archstarcoder2
·
parameters3.03B
·
quantizationQ4_0
1.7GB
params
{"stop":["<file_sep>","<|end_of_text|>"]}
41B
template
<file_sep>
{{- if .Suffix }}<fim_prefix>
{{ .Prompt }}<fim_suffix>{{ .Suffix }}<fim_middle>
{{- else
142B
license
BigCode Open RAIL-M v1 License Agreement
Section I: Preamble
This OpenRAIL-M License Agreement was
13kB
Readme
Supporting a context window of up to 16,384 tokens, StarCoder2 is the next generation of transparently trained open code LLMs.
starcoder2:instruct
: a 15B model that follows natural and human-written instructionsstarcoder2:15b
was trained on 600+ programming languages and 4+ trillion tokens.starcoder2:7b
was trained on 17 programming languages and 3.5+ trillion tokens.starcoder2:3b
was trained on 17 programming languages and 3+ trillion tokens.
StarCoder2-15B is the best in its size class and matches 33B+ models on many evaluations. StarCoder2-3B matches the performance of StarCoder1-15B.