The model used is a quantized version of `Llama-3-Taiwan-8B-Instruct`. More details can be found on the https://huggingface.co/yentinglin/Llama-3-Taiwan-8B-Instruct

611 4 months ago

Readme

The model used is a quantized version of Llama-3 Taiwan 8B Instruct, a specialized model designed for traditional Chinese conversation with 8 billion parameters. Quantization reduces the model’s size and computational requirements while maintaining performance, making it suitable for deployment in resource-constrained environments. More details can be found on the Hugging Face page.