Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
cwchang
/
llama-3-taiwan-70b-instruct
:q6_k
254
Downloads
Updated
1 year ago
The model used is a quantized version of `Llama-3-Taiwan-70B-Instruct`. More details can be found on the website (https://huggingface.co/yentinglin/Llama-3-Taiwan-70B-Instruct)
The model used is a quantized version of `Llama-3-Taiwan-70B-Instruct`. More details can be found on the website (https://huggingface.co/yentinglin/Llama-3-Taiwan-70B-Instruct)
Cancel
llama-3-taiwan-70b-instruct:q6_k
...
/
model
7b0e90f9e6ab · 10.0GB
Metadata
split.count
2
2
split.no
1
1
split.tensors.count
723
723
Tensor
Name
Type
Shape
blk.67
blk.67.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.67.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.67.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.67.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.67.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.67.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.67.ffn_norm.weight
F32
F32
[8192]
blk.67.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.68
blk.68.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.68.attn_norm.weight
F32
F32
[8192]
blk.68.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.68.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.68.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.68.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.68.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.68.ffn_norm.weight
F32
F32
[8192]
blk.68.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.69
blk.69.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.69.attn_norm.weight
F32
F32
[8192]
blk.69.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.69.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.69.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.69.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.69.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.69.ffn_norm.weight
F32
F32
[8192]
blk.69.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.70
blk.70.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.70.attn_norm.weight
F32
F32
[8192]
blk.70.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.70.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.70.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.70.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.70.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.70.ffn_norm.weight
F32
F32
[8192]
blk.70.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.71
blk.71.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.71.attn_norm.weight
F32
F32
[8192]
blk.71.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.71.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.71.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.71.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.71.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.71.ffn_norm.weight
F32
F32
[8192]
blk.71.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.72
blk.72.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.72.attn_norm.weight
F32
F32
[8192]
blk.72.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.72.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.72.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.72.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.72.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.72.ffn_norm.weight
F32
F32
[8192]
blk.72.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.73
blk.73.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.73.attn_norm.weight
F32
F32
[8192]
blk.73.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.73.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.73.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.73.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.73.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.73.ffn_norm.weight
F32
F32
[8192]
blk.73.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.74
blk.74.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.74.attn_norm.weight
F32
F32
[8192]
blk.74.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.74.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.74.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.74.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.74.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.74.ffn_norm.weight
F32
F32
[8192]
blk.74.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.75
blk.75.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.75.attn_norm.weight
F32
F32
[8192]
blk.75.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.75.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.75.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.75.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.75.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.75.ffn_norm.weight
F32
F32
[8192]
blk.75.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.76
blk.76.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.76.attn_norm.weight
F32
F32
[8192]
blk.76.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.76.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.76.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.76.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.76.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.76.ffn_norm.weight
F32
F32
[8192]
blk.76.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.77
blk.77.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.77.attn_norm.weight
F32
F32
[8192]
blk.77.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.77.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.77.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.77.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.77.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.77.ffn_norm.weight
F32
F32
[8192]
blk.77.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.78
blk.78.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.78.attn_norm.weight
F32
F32
[8192]
blk.78.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.78.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.78.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.78.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.78.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.78.ffn_norm.weight
F32
F32
[8192]
blk.78.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
blk.79
blk.79.attn_k.weight
Q6_K
Q6_K
[8192, 1024]
blk.79.attn_norm.weight
F32
F32
[8192]
blk.79.attn_output.weight
Q6_K
Q6_K
[8192, 8192]
blk.79.attn_q.weight
Q6_K
Q6_K
[8192, 8192]
blk.79.attn_v.weight
Q6_K
Q6_K
[8192, 1024]
blk.79.ffn_down.weight
Q6_K
Q6_K
[28672, 8192]
blk.79.ffn_gate.weight
Q6_K
Q6_K
[8192, 28672]
blk.79.ffn_norm.weight
F32
F32
[8192]
blk.79.ffn_up.weight
Q6_K
Q6_K
[8192, 28672]
output.weight
Q6_K
Q6_K
[8192, 128256]
output_norm.weight
F32
F32
[8192]