Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation.
2,007 Pulls Updated 6 weeks ago
d76121b83ea6 · 44MB
-
model.decoder.embed_positions.weightF32[384, 448]
-
model.decoder.embed_tokens.weightQ8_0[384, 51865]
-
model.decoder.layer_norm.biasF32[384]
-
model.decoder.layer_norm.weightF32[384]
-
model.decoder.layers.0.encoder_attn.k_proj.weightQ8_0[384, 384]
-
model.decoder.layers.0.encoder_attn.out_proj.biasF32[384]
-
model.decoder.layers.0.encoder_attn.out_proj.weightQ8_0[384, 384]
-
model.decoder.layers.0.encoder_attn.q_proj.biasF32[384]
-
model.decoder.layers.0.encoder_attn.q_proj.weightQ8_0[384, 384]
-
model.decoder.layers.0.encoder_attn.v_proj.biasF32[384]
-
model.decoder.layers.0.encoder_attn.v_proj.weightQ8_0[384, 384]
-
model.decoder.layers.0.encoder_attn_layer_norm.biasF32[384]
-
model.decoder.layers.0.encoder_attn_layer_norm.weightF32[384]
-
model.decoder.layers.0.fc1.biasF32[1536]
-
model.decoder.layers.0.fc1.weightQ8_0[384, 1536]
-
model.decoder.layers.0.fc2.biasF32[384]
-
model.decoder.layers.0.fc2.weightQ8_0[1536, 384]
-
model.decoder.layers.0.final_layer_norm.biasF32[384]
-
model.decoder.layers.0.final_layer_norm.weightF32[384]
-
model.decoder.layers.0.self_attn.k_proj.weightQ8_0[384, 384]
-
model.decoder.layers.0.self_attn.out_proj.biasF32[384]
-
model.decoder.layers.0.self_attn.out_proj.weightQ8_0[384, 384]
-
model.decoder.layers.0.self_attn.q_proj.biasF32[384]
-
model.decoder.layers.0.self_attn.q_proj.weightQ8_0[384, 384]
-
model.decoder.layers.0.self_attn.v_proj.biasF32[384]
-
model.decoder.layers.0.self_attn.v_proj.weightQ8_0[384, 384]
-
model.decoder.layers.0.self_attn_layer_norm.biasF32[384]
-
model.decoder.layers.0.self_attn_layer_norm.weightF32[384]
-
model.decoder.layers.1.encoder_attn.k_proj.weightQ8_0[384, 384]
-
model.decoder.layers.1.encoder_attn.out_proj.biasF32[384]
-
model.decoder.layers.1.encoder_attn.out_proj.weightQ8_0[384, 384]
-
model.decoder.layers.1.encoder_attn.q_proj.biasF32[384]
-
model.decoder.layers.1.encoder_attn.q_proj.weightQ8_0[384, 384]
-
model.decoder.layers.1.encoder_attn.v_proj.biasF32[384]
-
model.decoder.layers.1.encoder_attn.v_proj.weightQ8_0[384, 384]
-
model.decoder.layers.1.encoder_attn_layer_norm.biasF32[384]
-
model.decoder.layers.1.encoder_attn_layer_norm.weightF32[384]
-
model.decoder.layers.1.fc1.biasF32[1536]
-
model.decoder.layers.1.fc1.weightQ8_0[384, 1536]
-
model.decoder.layers.1.fc2.biasF32[384]
-
model.decoder.layers.1.fc2.weightQ8_0[1536, 384]
-
model.decoder.layers.1.final_layer_norm.biasF32[384]
-
model.decoder.layers.1.final_layer_norm.weightF32[384]
-
model.decoder.layers.1.self_attn.k_proj.weightQ8_0[384, 384]
-
model.decoder.layers.1.self_attn.out_proj.biasF32[384]
-
model.decoder.layers.1.self_attn.out_proj.weightQ8_0[384, 384]
-
model.decoder.layers.1.self_attn.q_proj.biasF32[384]
-
model.decoder.layers.1.self_attn.q_proj.weightQ8_0[384, 384]
-
model.decoder.layers.1.self_attn.v_proj.biasF32[384]
-
model.decoder.layers.1.self_attn.v_proj.weightQ8_0[384, 384]
-
model.decoder.layers.1.self_attn_layer_norm.biasF32[384]
-
model.decoder.layers.1.self_attn_layer_norm.weightF32[384]
-
model.decoder.layers.2.encoder_attn.k_proj.weightQ8_0[384, 384]
-
model.decoder.layers.2.encoder_attn.out_proj.biasF32[384]
-
model.decoder.layers.2.encoder_attn.out_proj.weightQ8_0[384, 384]
-
model.decoder.layers.2.encoder_attn.q_proj.biasF32[384]
-
model.decoder.layers.2.encoder_attn.q_proj.weightQ8_0[384, 384]
-
model.decoder.layers.2.encoder_attn.v_proj.biasF32[384]
-
model.decoder.layers.2.encoder_attn.v_proj.weightQ8_0[384, 384]
-
model.decoder.layers.2.encoder_attn_layer_norm.biasF32[384]
-
model.decoder.layers.2.encoder_attn_layer_norm.weightF32[384]
-
model.decoder.layers.2.fc1.biasF32[1536]
-
model.decoder.layers.2.fc1.weightQ8_0[384, 1536]
-
model.decoder.layers.2.fc2.biasF32[384]
-
model.decoder.layers.2.fc2.weightQ8_0[1536, 384]
-
model.decoder.layers.2.final_layer_norm.biasF32[384]
-
model.decoder.layers.2.final_layer_norm.weightF32[384]
-
model.decoder.layers.2.self_attn.k_proj.weightQ8_0[384, 384]
-
model.decoder.layers.2.self_attn.out_proj.biasF32[384]
-
model.decoder.layers.2.self_attn.out_proj.weightQ8_0[384, 384]
-
model.decoder.layers.2.self_attn.q_proj.biasF32[384]
-
model.decoder.layers.2.self_attn.q_proj.weightQ8_0[384, 384]
-
model.decoder.layers.2.self_attn.v_proj.biasF32[384]
-
model.decoder.layers.2.self_attn.v_proj.weightQ8_0[384, 384]
-
model.decoder.layers.2.self_attn_layer_norm.biasF32[384]
-
model.decoder.layers.2.self_attn_layer_norm.weightF32[384]
-
model.decoder.layers.3.encoder_attn.k_proj.weightQ8_0[384, 384]
-
model.decoder.layers.3.encoder_attn.out_proj.biasF32[384]
-
model.decoder.layers.3.encoder_attn.out_proj.weightQ8_0[384, 384]
-
model.decoder.layers.3.encoder_attn.q_proj.biasF32[384]
-
model.decoder.layers.3.encoder_attn.q_proj.weightQ8_0[384, 384]
-
model.decoder.layers.3.encoder_attn.v_proj.biasF32[384]
-
model.decoder.layers.3.encoder_attn.v_proj.weightQ8_0[384, 384]
-
model.decoder.layers.3.encoder_attn_layer_norm.biasF32[384]
-
model.decoder.layers.3.encoder_attn_layer_norm.weightF32[384]
-
model.decoder.layers.3.fc1.biasF32[1536]
-
model.decoder.layers.3.fc1.weightQ8_0[384, 1536]
-
model.decoder.layers.3.fc2.biasF32[384]
-
model.decoder.layers.3.fc2.weightQ8_0[1536, 384]
-
model.decoder.layers.3.final_layer_norm.biasF32[384]
-
model.decoder.layers.3.final_layer_norm.weightF32[384]
-
model.decoder.layers.3.self_attn.k_proj.weightQ8_0[384, 384]
-
model.decoder.layers.3.self_attn.out_proj.biasF32[384]
-
model.decoder.layers.3.self_attn.out_proj.weightQ8_0[384, 384]
-
model.decoder.layers.3.self_attn.q_proj.biasF32[384]
-
model.decoder.layers.3.self_attn.q_proj.weightQ8_0[384, 384]
-
model.decoder.layers.3.self_attn.v_proj.biasF32[384]
-
model.decoder.layers.3.self_attn.v_proj.weightQ8_0[384, 384]
-
model.decoder.layers.3.self_attn_layer_norm.biasF32[384]
-
model.decoder.layers.3.self_attn_layer_norm.weightF32[384]
-
model.encoder.conv1.biasF32[384]
-
model.encoder.conv1.weightF32[3, 80, 384]
-
model.encoder.conv2.biasF32[384]
-
model.encoder.conv2.weightF32[3, 384, 384]
-
model.encoder.embed_positions.weightF32[384, 1500]
-
model.encoder.layer_norm.biasF32[384]
-
model.encoder.layer_norm.weightF32[384]
-
model.encoder.layers.0.fc1.biasF32[1536]
-
model.encoder.layers.0.fc1.weightQ8_0[384, 1536]
-
model.encoder.layers.0.fc2.biasF32[384]
-
model.encoder.layers.0.fc2.weightQ8_0[1536, 384]
-
model.encoder.layers.0.final_layer_norm.biasF32[384]
-
model.encoder.layers.0.final_layer_norm.weightF32[384]
-
model.encoder.layers.0.self_attn.k_proj.weightQ8_0[384, 384]
-
model.encoder.layers.0.self_attn.out_proj.biasF32[384]
-
model.encoder.layers.0.self_attn.out_proj.weightQ8_0[384, 384]
-
model.encoder.layers.0.self_attn.q_proj.biasF32[384]
-
model.encoder.layers.0.self_attn.q_proj.weightQ8_0[384, 384]
-
model.encoder.layers.0.self_attn.v_proj.biasF32[384]
-
model.encoder.layers.0.self_attn.v_proj.weightQ8_0[384, 384]
-
model.encoder.layers.0.self_attn_layer_norm.biasF32[384]
-
model.encoder.layers.0.self_attn_layer_norm.weightF32[384]
-
model.encoder.layers.1.fc1.biasF32[1536]
-
model.encoder.layers.1.fc1.weightQ8_0[384, 1536]
-
model.encoder.layers.1.fc2.biasF32[384]
-
model.encoder.layers.1.fc2.weightQ8_0[1536, 384]
-
model.encoder.layers.1.final_layer_norm.biasF32[384]
-
model.encoder.layers.1.final_layer_norm.weightF32[384]
-
model.encoder.layers.1.self_attn.k_proj.weightQ8_0[384, 384]
-
model.encoder.layers.1.self_attn.out_proj.biasF32[384]
-
model.encoder.layers.1.self_attn.out_proj.weightQ8_0[384, 384]
-
model.encoder.layers.1.self_attn.q_proj.biasF32[384]
-
model.encoder.layers.1.self_attn.q_proj.weightQ8_0[384, 384]
-
model.encoder.layers.1.self_attn.v_proj.biasF32[384]
-
model.encoder.layers.1.self_attn.v_proj.weightQ8_0[384, 384]
-
model.encoder.layers.1.self_attn_layer_norm.biasF32[384]
-
model.encoder.layers.1.self_attn_layer_norm.weightF32[384]
-
model.encoder.layers.2.fc1.biasF32[1536]
-
model.encoder.layers.2.fc1.weightQ8_0[384, 1536]
-
model.encoder.layers.2.fc2.biasF32[384]
-
model.encoder.layers.2.fc2.weightQ8_0[1536, 384]
-
model.encoder.layers.2.final_layer_norm.biasF32[384]
-
model.encoder.layers.2.final_layer_norm.weightF32[384]
-
model.encoder.layers.2.self_attn.k_proj.weightQ8_0[384, 384]
-
model.encoder.layers.2.self_attn.out_proj.biasF32[384]
-
model.encoder.layers.2.self_attn.out_proj.weightQ8_0[384, 384]
-
model.encoder.layers.2.self_attn.q_proj.biasF32[384]
-
model.encoder.layers.2.self_attn.q_proj.weightQ8_0[384, 384]
-
model.encoder.layers.2.self_attn.v_proj.biasF32[384]
-
model.encoder.layers.2.self_attn.v_proj.weightQ8_0[384, 384]
-
model.encoder.layers.2.self_attn_layer_norm.biasF32[384]
-
model.encoder.layers.2.self_attn_layer_norm.weightF32[384]
-
model.encoder.layers.3.fc1.biasF32[1536]
-
model.encoder.layers.3.fc1.weightQ8_0[384, 1536]
-
model.encoder.layers.3.fc2.biasF32[384]
-
model.encoder.layers.3.fc2.weightQ8_0[1536, 384]
-
model.encoder.layers.3.final_layer_norm.biasF32[384]
-
model.encoder.layers.3.final_layer_norm.weightF32[384]
-
model.encoder.layers.3.self_attn.k_proj.weightQ8_0[384, 384]
-
model.encoder.layers.3.self_attn.out_proj.biasF32[384]
-
model.encoder.layers.3.self_attn.out_proj.weightQ8_0[384, 384]
-
model.encoder.layers.3.self_attn.q_proj.biasF32[384]
-
model.encoder.layers.3.self_attn.q_proj.weightQ8_0[384, 384]
-
model.encoder.layers.3.self_attn.v_proj.biasF32[384]
-
model.encoder.layers.3.self_attn.v_proj.weightQ8_0[384, 384]
-
model.encoder.layers.3.self_attn_layer_norm.biasF32[384]
-
model.encoder.layers.3.self_attn_layer_norm.weightF32[384]
Metadata
Tensor