Phi-3 128K version of Q5
12.4K Pulls Updated 8 months ago
Updated 8 months ago
8 months ago
d8f0cff3daa1 · 2.7GB
model
archllama
·
parameters3.82B
·
quantizationQ5_K_M
2.7GB
params
{
"stop": [
"<|user|>",
"<|assistant|>",
"<|system|>",
"<|end|>"
127B
template
{{ if .System }}<|system|>
{{ .System }}<|end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|end
149B
Readme
Convert from PrunaAI/Phi-3-mini-128k-instruct-GGUF-Imatrix-smashed, adopt Q5_K_8_4 quantization.
Its multilingual capabilities are clearly superior to version Q4 quantization.