The model used is a quantized version of `Llama-3-Taiwan-8B-Instruct`. More details can be found on the https://huggingface.co/yentinglin/Llama-3-Taiwan-8B-Instruct
1,525 Pulls Updated 8 months ago
Updated 9 months ago
9 months ago
3f0396cc71cc · 5.7GB
model
archllama
·
parameters8.03B
·
quantizationQ5_K_M
5.7GB
params
{
"num_ctx": 8192,
"stop": [
"<|start_header_id|>",
"<|end_header_id|>",
171B
template
"{{ if .System }}<|start_header_id|>system<|end_header_id|>
{{ .System }}<|eot_id|>{{ end }}{{ if .
257B
Readme
The model used is a quantized version of Llama-3 Taiwan 8B Instruct, a specialized model designed for traditional Chinese conversation with 8 billion parameters. Quantization reduces the model’s size and computational requirements while maintaining performance, making it suitable for deployment in resource-constrained environments. More details can be found on the Hugging Face page.