Models
GitHub
Discord
Turbo
Sign in
Download
Models
Download
GitHub
Discord
Sign in
dimavz
/
whisper-tiny
:latest
7,512
Downloads
Updated
5 months ago
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation.
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation.
Cancel
whisper-tiny:latest
...
/
model
d76121b83ea6 · 44MB
Metadata
Tensor
Name
Type
Shape
model.decoder.embed_positions.weight
F32
F32
[384, 448]
model.decoder.embed_tokens.weight
Q8_0
Q8_0
[384, 51865]
model.decoder.layer_norm.bias
F32
F32
[384]
model.decoder.layer_norm.weight
F32
F32
[384]
model.decoder.layers.0.encoder_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.0.encoder_attn.out_proj.bias
F32
F32
[384]
model.decoder.layers.0.encoder_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.0.encoder_attn.q_proj.bias
F32
F32
[384]
model.decoder.layers.0.encoder_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.0.encoder_attn.v_proj.bias
F32
F32
[384]
model.decoder.layers.0.encoder_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.0.encoder_attn_layer_norm.bias
F32
F32
[384]
model.decoder.layers.0.encoder_attn_layer_norm.weight
F32
F32
[384]
model.decoder.layers.0.fc1.bias
F32
F32
[1536]
model.decoder.layers.0.fc1.weight
Q8_0
Q8_0
[384, 1536]
model.decoder.layers.0.fc2.bias
F32
F32
[384]
model.decoder.layers.0.fc2.weight
Q8_0
Q8_0
[1536, 384]
model.decoder.layers.0.final_layer_norm.bias
F32
F32
[384]
model.decoder.layers.0.final_layer_norm.weight
F32
F32
[384]
model.decoder.layers.0.self_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.0.self_attn.out_proj.bias
F32
F32
[384]
model.decoder.layers.0.self_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.0.self_attn.q_proj.bias
F32
F32
[384]
model.decoder.layers.0.self_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.0.self_attn.v_proj.bias
F32
F32
[384]
model.decoder.layers.0.self_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.0.self_attn_layer_norm.bias
F32
F32
[384]
model.decoder.layers.0.self_attn_layer_norm.weight
F32
F32
[384]
model.decoder.layers.1.encoder_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.1.encoder_attn.out_proj.bias
F32
F32
[384]
model.decoder.layers.1.encoder_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.1.encoder_attn.q_proj.bias
F32
F32
[384]
model.decoder.layers.1.encoder_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.1.encoder_attn.v_proj.bias
F32
F32
[384]
model.decoder.layers.1.encoder_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.1.encoder_attn_layer_norm.bias
F32
F32
[384]
model.decoder.layers.1.encoder_attn_layer_norm.weight
F32
F32
[384]
model.decoder.layers.1.fc1.bias
F32
F32
[1536]
model.decoder.layers.1.fc1.weight
Q8_0
Q8_0
[384, 1536]
model.decoder.layers.1.fc2.bias
F32
F32
[384]
model.decoder.layers.1.fc2.weight
Q8_0
Q8_0
[1536, 384]
model.decoder.layers.1.final_layer_norm.bias
F32
F32
[384]
model.decoder.layers.1.final_layer_norm.weight
F32
F32
[384]
model.decoder.layers.1.self_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.1.self_attn.out_proj.bias
F32
F32
[384]
model.decoder.layers.1.self_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.1.self_attn.q_proj.bias
F32
F32
[384]
model.decoder.layers.1.self_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.1.self_attn.v_proj.bias
F32
F32
[384]
model.decoder.layers.1.self_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.1.self_attn_layer_norm.bias
F32
F32
[384]
model.decoder.layers.1.self_attn_layer_norm.weight
F32
F32
[384]
model.decoder.layers.2.encoder_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.2.encoder_attn.out_proj.bias
F32
F32
[384]
model.decoder.layers.2.encoder_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.2.encoder_attn.q_proj.bias
F32
F32
[384]
model.decoder.layers.2.encoder_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.2.encoder_attn.v_proj.bias
F32
F32
[384]
model.decoder.layers.2.encoder_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.2.encoder_attn_layer_norm.bias
F32
F32
[384]
model.decoder.layers.2.encoder_attn_layer_norm.weight
F32
F32
[384]
model.decoder.layers.2.fc1.bias
F32
F32
[1536]
model.decoder.layers.2.fc1.weight
Q8_0
Q8_0
[384, 1536]
model.decoder.layers.2.fc2.bias
F32
F32
[384]
model.decoder.layers.2.fc2.weight
Q8_0
Q8_0
[1536, 384]
model.decoder.layers.2.final_layer_norm.bias
F32
F32
[384]
model.decoder.layers.2.final_layer_norm.weight
F32
F32
[384]
model.decoder.layers.2.self_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.2.self_attn.out_proj.bias
F32
F32
[384]
model.decoder.layers.2.self_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.2.self_attn.q_proj.bias
F32
F32
[384]
model.decoder.layers.2.self_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.2.self_attn.v_proj.bias
F32
F32
[384]
model.decoder.layers.2.self_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.2.self_attn_layer_norm.bias
F32
F32
[384]
model.decoder.layers.2.self_attn_layer_norm.weight
F32
F32
[384]
model.decoder.layers.3.encoder_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.3.encoder_attn.out_proj.bias
F32
F32
[384]
model.decoder.layers.3.encoder_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.3.encoder_attn.q_proj.bias
F32
F32
[384]
model.decoder.layers.3.encoder_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.3.encoder_attn.v_proj.bias
F32
F32
[384]
model.decoder.layers.3.encoder_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.3.encoder_attn_layer_norm.bias
F32
F32
[384]
model.decoder.layers.3.encoder_attn_layer_norm.weight
F32
F32
[384]
model.decoder.layers.3.fc1.bias
F32
F32
[1536]
model.decoder.layers.3.fc1.weight
Q8_0
Q8_0
[384, 1536]
model.decoder.layers.3.fc2.bias
F32
F32
[384]
model.decoder.layers.3.fc2.weight
Q8_0
Q8_0
[1536, 384]
model.decoder.layers.3.final_layer_norm.bias
F32
F32
[384]
model.decoder.layers.3.final_layer_norm.weight
F32
F32
[384]
model.decoder.layers.3.self_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.3.self_attn.out_proj.bias
F32
F32
[384]
model.decoder.layers.3.self_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.3.self_attn.q_proj.bias
F32
F32
[384]
model.decoder.layers.3.self_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.3.self_attn.v_proj.bias
F32
F32
[384]
model.decoder.layers.3.self_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.decoder.layers.3.self_attn_layer_norm.bias
F32
F32
[384]
model.decoder.layers.3.self_attn_layer_norm.weight
F32
F32
[384]
model.encoder.conv1.bias
F32
F32
[384]
model.encoder.conv1.weight
F32
F32
[3, 80, 384]
model.encoder.conv2.bias
F32
F32
[384]
model.encoder.conv2.weight
F32
F32
[3, 384, 384]
model.encoder.embed_positions.weight
F32
F32
[384, 1500]
model.encoder.layer_norm.bias
F32
F32
[384]
model.encoder.layer_norm.weight
F32
F32
[384]
model.encoder.layers.0.fc1.bias
F32
F32
[1536]
model.encoder.layers.0.fc1.weight
Q8_0
Q8_0
[384, 1536]
model.encoder.layers.0.fc2.bias
F32
F32
[384]
model.encoder.layers.0.fc2.weight
Q8_0
Q8_0
[1536, 384]
model.encoder.layers.0.final_layer_norm.bias
F32
F32
[384]
model.encoder.layers.0.final_layer_norm.weight
F32
F32
[384]
model.encoder.layers.0.self_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.0.self_attn.out_proj.bias
F32
F32
[384]
model.encoder.layers.0.self_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.0.self_attn.q_proj.bias
F32
F32
[384]
model.encoder.layers.0.self_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.0.self_attn.v_proj.bias
F32
F32
[384]
model.encoder.layers.0.self_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.0.self_attn_layer_norm.bias
F32
F32
[384]
model.encoder.layers.0.self_attn_layer_norm.weight
F32
F32
[384]
model.encoder.layers.1.fc1.bias
F32
F32
[1536]
model.encoder.layers.1.fc1.weight
Q8_0
Q8_0
[384, 1536]
model.encoder.layers.1.fc2.bias
F32
F32
[384]
model.encoder.layers.1.fc2.weight
Q8_0
Q8_0
[1536, 384]
model.encoder.layers.1.final_layer_norm.bias
F32
F32
[384]
model.encoder.layers.1.final_layer_norm.weight
F32
F32
[384]
model.encoder.layers.1.self_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.1.self_attn.out_proj.bias
F32
F32
[384]
model.encoder.layers.1.self_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.1.self_attn.q_proj.bias
F32
F32
[384]
model.encoder.layers.1.self_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.1.self_attn.v_proj.bias
F32
F32
[384]
model.encoder.layers.1.self_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.1.self_attn_layer_norm.bias
F32
F32
[384]
model.encoder.layers.1.self_attn_layer_norm.weight
F32
F32
[384]
model.encoder.layers.2.fc1.bias
F32
F32
[1536]
model.encoder.layers.2.fc1.weight
Q8_0
Q8_0
[384, 1536]
model.encoder.layers.2.fc2.bias
F32
F32
[384]
model.encoder.layers.2.fc2.weight
Q8_0
Q8_0
[1536, 384]
model.encoder.layers.2.final_layer_norm.bias
F32
F32
[384]
model.encoder.layers.2.final_layer_norm.weight
F32
F32
[384]
model.encoder.layers.2.self_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.2.self_attn.out_proj.bias
F32
F32
[384]
model.encoder.layers.2.self_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.2.self_attn.q_proj.bias
F32
F32
[384]
model.encoder.layers.2.self_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.2.self_attn.v_proj.bias
F32
F32
[384]
model.encoder.layers.2.self_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.2.self_attn_layer_norm.bias
F32
F32
[384]
model.encoder.layers.2.self_attn_layer_norm.weight
F32
F32
[384]
model.encoder.layers.3.fc1.bias
F32
F32
[1536]
model.encoder.layers.3.fc1.weight
Q8_0
Q8_0
[384, 1536]
model.encoder.layers.3.fc2.bias
F32
F32
[384]
model.encoder.layers.3.fc2.weight
Q8_0
Q8_0
[1536, 384]
model.encoder.layers.3.final_layer_norm.bias
F32
F32
[384]
model.encoder.layers.3.final_layer_norm.weight
F32
F32
[384]
model.encoder.layers.3.self_attn.k_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.3.self_attn.out_proj.bias
F32
F32
[384]
model.encoder.layers.3.self_attn.out_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.3.self_attn.q_proj.bias
F32
F32
[384]
model.encoder.layers.3.self_attn.q_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.3.self_attn.v_proj.bias
F32
F32
[384]
model.encoder.layers.3.self_attn.v_proj.weight
Q8_0
Q8_0
[384, 384]
model.encoder.layers.3.self_attn_layer_norm.bias
F32
F32
[384]
model.encoder.layers.3.self_attn_layer_norm.weight
F32
F32
[384]