2,091 Downloads Updated 1 year ago
Name
17 models
llama-3-taiwan-8b-instruct:latest
5.7GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q2_k
3.2GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q3_k_s
3.7GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q3_k_m
4.0GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q3_k_l
4.3GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q4_0
4.7GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q4_1
5.1GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q4_k_s
4.7GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q4_k_m
4.9GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q5_0
5.6GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q5_1
6.1GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q5_k_s
5.6GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q5_k_m
5.7GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q6_k
6.6GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:q8_0
8.5GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:f16
16GB · 8K context window · Text · 1 year ago
llama-3-taiwan-8b-instruct:f32
32GB · 8K context window · Text · 1 year ago
The model used is a quantized version of Llama-3 Taiwan 8B Instruct, a specialized model designed for traditional Chinese conversation with 8 billion parameters. Quantization reduces the model’s size and computational requirements while maintaining performance, making it suitable for deployment in resource-constrained environments. More details can be found on the Hugging Face page.