Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
jina
/
jina-embeddings-v2-base-es
:latest
16.1K
Downloads
Updated
1 year ago
Text embedding model (base) for English and Spanish input of size up to 8192 tokens
Text embedding model (base) for English and Spanish input of size up to 8192 tokens
Cancel
embedding
jina-embeddings-v2-base-es:latest
...
/
model
026d225ac973 · 323MB
Metadata
general.architecture
jina-bert-v2
jina-bert-v2
general.file_type
F16
F16
jina-bert-v2.attention.causal
false
false
jina-bert-v2.attention.head_count
12
12
jina-bert-v2.attention.layer_norm_epsilon
1e-12
1e-12
jina-bert-v2.block_count
12
12
jina-bert-v2.context_length
8192
8192
jina-bert-v2.embedding_length
768
768
jina-bert-v2.feed_forward_length
3072
3072
jina-bert-v2.pooling_type
Mean
Mean
tokenizer.ggml.add_bos_token
true
true
tokenizer.ggml.add_eos_token
true
true
tokenizer.ggml.bos_token_id
0
0
tokenizer.ggml.cls_token_id
0
0
tokenizer.ggml.eos_token_id
2
2
tokenizer.ggml.mask_token_id
4
4
tokenizer.ggml.merges
[Ġ a, Ġ t, e n, e s, e r, ...]
[Ġ a, Ġ t, e n, e s, e r, ...]
tokenizer.ggml.model
gpt2
gpt2
tokenizer.ggml.padding_token_id
1
1
tokenizer.ggml.pre
jina-v2-es
jina-v2-es
tokenizer.ggml.seperator_token_id
2
2
tokenizer.ggml.token_type
[3, 3, 3, 3, 3, ...]
[3, 3, 3, 3, 3, ...]
tokenizer.ggml.token_type_count
2
2
tokenizer.ggml.tokens
[<s>, <pad>, </s>, <unk>, <mask>, ...]
[<s>, <pad>, </s>, <unk>, <mask>, ...]
tokenizer.ggml.unknown_token_id
3
3
Tensor
Name
Type
Shape
token_embd.weight
F16
F16
[768, 61056]
blk.0
blk.0.attn_k.bias
F32
F32
[768]
blk.0.attn_k.weight
F16
F16
[768, 768]
blk.0.attn_k_norm.bias
F32
F32
[768]
blk.0.attn_k_norm.weight
F32
F32
[768]
blk.0.attn_output.bias
F32
F32
[768]
blk.0.attn_output.weight
F16
F16
[768, 768]
blk.0.attn_output_norm.bias
F32
F32
[768]
blk.0.attn_output_norm.weight
F32
F32
[768]
blk.0.attn_q.bias
F32
F32
[768]
blk.0.attn_q.weight
F16
F16
[768, 768]
blk.0.attn_q_norm.bias
F32
F32
[768]
blk.0.attn_q_norm.weight
F32
F32
[768]
blk.0.attn_v.bias
F32
F32
[768]
blk.0.attn_v.weight
F16
F16
[768, 768]
blk.0.ffn_down.bias
F32
F32
[768]
blk.0.ffn_down.weight
F16
F16
[3072, 768]
blk.0.ffn_gate.weight
F16
F16
[768, 3072]
blk.0.ffn_up.weight
F16
F16
[768, 3072]
blk.0.layer_output_norm.bias
F32
F32
[768]
blk.0.layer_output_norm.weight
F32
F32
[768]
blk.1
blk.1.attn_k.bias
F32
F32
[768]
blk.1.attn_k.weight
F16
F16
[768, 768]
blk.1.attn_k_norm.bias
F32
F32
[768]
blk.1.attn_k_norm.weight
F32
F32
[768]
blk.1.attn_output.bias
F32
F32
[768]
blk.1.attn_output.weight
F16
F16
[768, 768]
blk.1.attn_output_norm.bias
F32
F32
[768]
blk.1.attn_output_norm.weight
F32
F32
[768]
blk.1.attn_q.bias
F32
F32
[768]
blk.1.attn_q.weight
F16
F16
[768, 768]
blk.1.attn_q_norm.bias
F32
F32
[768]
blk.1.attn_q_norm.weight
F32
F32
[768]
blk.1.attn_v.bias
F32
F32
[768]
blk.1.attn_v.weight
F16
F16
[768, 768]
blk.1.ffn_down.bias
F32
F32
[768]
blk.1.ffn_down.weight
F16
F16
[3072, 768]
blk.1.ffn_gate.weight
F16
F16
[768, 3072]
blk.1.ffn_up.weight
F16
F16
[768, 3072]
blk.1.layer_output_norm.bias
F32
F32
[768]
blk.1.layer_output_norm.weight
F32
F32
[768]
blk.2
blk.2.attn_k.bias
F32
F32
[768]
blk.2.attn_k.weight
F16
F16
[768, 768]
blk.2.attn_k_norm.bias
F32
F32
[768]
blk.2.attn_k_norm.weight
F32
F32
[768]
blk.2.attn_output.bias
F32
F32
[768]
blk.2.attn_output.weight
F16
F16
[768, 768]
blk.2.attn_output_norm.bias
F32
F32
[768]
blk.2.attn_output_norm.weight
F32
F32
[768]
blk.2.attn_q.bias
F32
F32
[768]
blk.2.attn_q.weight
F16
F16
[768, 768]
blk.2.attn_q_norm.bias
F32
F32
[768]
blk.2.attn_q_norm.weight
F32
F32
[768]
blk.2.attn_v.bias
F32
F32
[768]
blk.2.attn_v.weight
F16
F16
[768, 768]
blk.2.ffn_down.bias
F32
F32
[768]
blk.2.ffn_down.weight
F16
F16
[3072, 768]
blk.2.ffn_gate.weight
F16
F16
[768, 3072]
blk.2.ffn_up.weight
F16
F16
[768, 3072]
blk.2.layer_output_norm.bias
F32
F32
[768]
blk.2.layer_output_norm.weight
F32
F32
[768]
blk.3
blk.3.attn_k.bias
F32
F32
[768]
blk.3.attn_k.weight
F16
F16
[768, 768]
blk.3.attn_k_norm.bias
F32
F32
[768]
blk.3.attn_k_norm.weight
F32
F32
[768]
blk.3.attn_output.bias
F32
F32
[768]
blk.3.attn_output.weight
F16
F16
[768, 768]
blk.3.attn_output_norm.bias
F32
F32
[768]
blk.3.attn_output_norm.weight
F32
F32
[768]
blk.3.attn_q.bias
F32
F32
[768]
blk.3.attn_q.weight
F16
F16
[768, 768]
blk.3.attn_q_norm.bias
F32
F32
[768]
blk.3.attn_q_norm.weight
F32
F32
[768]
blk.3.attn_v.bias
F32
F32
[768]
blk.3.attn_v.weight
F16
F16
[768, 768]
blk.3.ffn_down.bias
F32
F32
[768]
blk.3.ffn_down.weight
F16
F16
[3072, 768]
blk.3.ffn_gate.weight
F16
F16
[768, 3072]
blk.3.ffn_up.weight
F16
F16
[768, 3072]
blk.3.layer_output_norm.bias
F32
F32
[768]
blk.3.layer_output_norm.weight
F32
F32
[768]
blk.4
blk.4.attn_k.bias
F32
F32
[768]
blk.4.attn_k.weight
F16
F16
[768, 768]
blk.4.attn_k_norm.bias
F32
F32
[768]
blk.4.attn_k_norm.weight
F32
F32
[768]
blk.4.attn_output.bias
F32
F32
[768]
blk.4.attn_output.weight
F16
F16
[768, 768]
blk.4.attn_output_norm.bias
F32
F32
[768]
blk.4.attn_output_norm.weight
F32
F32
[768]
blk.4.attn_q.bias
F32
F32
[768]
blk.4.attn_q.weight
F16
F16
[768, 768]
blk.4.attn_q_norm.bias
F32
F32
[768]
blk.4.attn_q_norm.weight
F32
F32
[768]
blk.4.attn_v.bias
F32
F32
[768]
blk.4.attn_v.weight
F16
F16
[768, 768]
blk.4.ffn_down.bias
F32
F32
[768]
blk.4.ffn_down.weight
F16
F16
[3072, 768]
blk.4.ffn_gate.weight
F16
F16
[768, 3072]
blk.4.ffn_up.weight
F16
F16
[768, 3072]
blk.4.layer_output_norm.bias
F32
F32
[768]
blk.4.layer_output_norm.weight
F32
F32
[768]
blk.5
blk.5.attn_k.bias
F32
F32
[768]
blk.5.attn_k.weight
F16
F16
[768, 768]
blk.5.attn_k_norm.bias
F32
F32
[768]
blk.5.attn_k_norm.weight
F32
F32
[768]
blk.5.attn_output.bias
F32
F32
[768]
blk.5.attn_output.weight
F16
F16
[768, 768]
blk.5.attn_output_norm.bias
F32
F32
[768]
blk.5.attn_output_norm.weight
F32
F32
[768]
blk.5.attn_q.bias
F32
F32
[768]
blk.5.attn_q.weight
F16
F16
[768, 768]
blk.5.attn_q_norm.bias
F32
F32
[768]
blk.5.attn_q_norm.weight
F32
F32
[768]
blk.5.attn_v.bias
F32
F32
[768]
blk.5.attn_v.weight
F16
F16
[768, 768]
blk.5.ffn_down.bias
F32
F32
[768]
blk.5.ffn_down.weight
F16
F16
[3072, 768]
blk.5.ffn_gate.weight
F16
F16
[768, 3072]
blk.5.ffn_up.weight
F16
F16
[768, 3072]
blk.5.layer_output_norm.bias
F32
F32
[768]
blk.5.layer_output_norm.weight
F32
F32
[768]
blk.6
blk.6.attn_k.bias
F32
F32
[768]
blk.6.attn_k.weight
F16
F16
[768, 768]
blk.6.attn_k_norm.bias
F32
F32
[768]
blk.6.attn_k_norm.weight
F32
F32
[768]
blk.6.attn_output.bias
F32
F32
[768]
blk.6.attn_output.weight
F16
F16
[768, 768]
blk.6.attn_output_norm.bias
F32
F32
[768]
blk.6.attn_output_norm.weight
F32
F32
[768]
blk.6.attn_q.bias
F32
F32
[768]
blk.6.attn_q.weight
F16
F16
[768, 768]
blk.6.attn_q_norm.bias
F32
F32
[768]
blk.6.attn_q_norm.weight
F32
F32
[768]
blk.6.attn_v.bias
F32
F32
[768]
blk.6.attn_v.weight
F16
F16
[768, 768]
blk.6.ffn_down.bias
F32
F32
[768]
blk.6.ffn_down.weight
F16
F16
[3072, 768]
blk.6.ffn_gate.weight
F16
F16
[768, 3072]
blk.6.ffn_up.weight
F16
F16
[768, 3072]
blk.6.layer_output_norm.bias
F32
F32
[768]
blk.6.layer_output_norm.weight
F32
F32
[768]
blk.7
blk.7.attn_k.bias
F32
F32
[768]
blk.7.attn_k.weight
F16
F16
[768, 768]
blk.7.attn_k_norm.bias
F32
F32
[768]
blk.7.attn_k_norm.weight
F32
F32
[768]
blk.7.attn_output.bias
F32
F32
[768]
blk.7.attn_output.weight
F16
F16
[768, 768]
blk.7.attn_output_norm.bias
F32
F32
[768]
blk.7.attn_output_norm.weight
F32
F32
[768]
blk.7.attn_q.bias
F32
F32
[768]
blk.7.attn_q.weight
F16
F16
[768, 768]
blk.7.attn_q_norm.bias
F32
F32
[768]
blk.7.attn_q_norm.weight
F32
F32
[768]
blk.7.attn_v.bias
F32
F32
[768]
blk.7.attn_v.weight
F16
F16
[768, 768]
blk.7.ffn_down.bias
F32
F32
[768]
blk.7.ffn_down.weight
F16
F16
[3072, 768]
blk.7.ffn_gate.weight
F16
F16
[768, 3072]
blk.7.ffn_up.weight
F16
F16
[768, 3072]
blk.7.layer_output_norm.bias
F32
F32
[768]
blk.7.layer_output_norm.weight
F32
F32
[768]
blk.8
blk.8.attn_k.bias
F32
F32
[768]
blk.8.attn_k.weight
F16
F16
[768, 768]
blk.8.attn_k_norm.bias
F32
F32
[768]
blk.8.attn_k_norm.weight
F32
F32
[768]
blk.8.attn_output.bias
F32
F32
[768]
blk.8.attn_output.weight
F16
F16
[768, 768]
blk.8.attn_output_norm.bias
F32
F32
[768]
blk.8.attn_output_norm.weight
F32
F32
[768]
blk.8.attn_q.bias
F32
F32
[768]
blk.8.attn_q.weight
F16
F16
[768, 768]
blk.8.attn_q_norm.bias
F32
F32
[768]
blk.8.attn_q_norm.weight
F32
F32
[768]
blk.8.attn_v.bias
F32
F32
[768]
blk.8.attn_v.weight
F16
F16
[768, 768]
blk.8.ffn_down.bias
F32
F32
[768]
blk.8.ffn_down.weight
F16
F16
[3072, 768]
blk.8.ffn_gate.weight
F16
F16
[768, 3072]
blk.8.ffn_up.weight
F16
F16
[768, 3072]
blk.8.layer_output_norm.bias
F32
F32
[768]
blk.8.layer_output_norm.weight
F32
F32
[768]
blk.9
blk.9.attn_k.bias
F32
F32
[768]
blk.9.attn_k.weight
F16
F16
[768, 768]
blk.9.attn_k_norm.bias
F32
F32
[768]
blk.9.attn_k_norm.weight
F32
F32
[768]
blk.9.attn_output.bias
F32
F32
[768]
blk.9.attn_output.weight
F16
F16
[768, 768]
blk.9.attn_output_norm.bias
F32
F32
[768]
blk.9.attn_output_norm.weight
F32
F32
[768]
blk.9.attn_q.bias
F32
F32
[768]
blk.9.attn_q.weight
F16
F16
[768, 768]
blk.9.attn_q_norm.bias
F32
F32
[768]
blk.9.attn_q_norm.weight
F32
F32
[768]
blk.9.attn_v.bias
F32
F32
[768]
blk.9.attn_v.weight
F16
F16
[768, 768]
blk.9.ffn_down.bias
F32
F32
[768]
blk.9.ffn_down.weight
F16
F16
[3072, 768]
blk.9.ffn_gate.weight
F16
F16
[768, 3072]
blk.9.ffn_up.weight
F16
F16
[768, 3072]
blk.9.layer_output_norm.bias
F32
F32
[768]
blk.9.layer_output_norm.weight
F32
F32
[768]
blk.10
blk.10.attn_k.bias
F32
F32
[768]
blk.10.attn_k.weight
F16
F16
[768, 768]
blk.10.attn_k_norm.bias
F32
F32
[768]
blk.10.attn_k_norm.weight
F32
F32
[768]
blk.10.attn_output.bias
F32
F32
[768]
blk.10.attn_output.weight
F16
F16
[768, 768]
blk.10.attn_output_norm.bias
F32
F32
[768]
blk.10.attn_output_norm.weight
F32
F32
[768]
blk.10.attn_q.bias
F32
F32
[768]
blk.10.attn_q.weight
F16
F16
[768, 768]
blk.10.attn_q_norm.bias
F32
F32
[768]
blk.10.attn_q_norm.weight
F32
F32
[768]
blk.10.attn_v.bias
F32
F32
[768]
blk.10.attn_v.weight
F16
F16
[768, 768]
blk.10.ffn_down.bias
F32
F32
[768]
blk.10.ffn_down.weight
F16
F16
[3072, 768]
blk.10.ffn_gate.weight
F16
F16
[768, 3072]
blk.10.ffn_up.weight
F16
F16
[768, 3072]
blk.10.layer_output_norm.bias
F32
F32
[768]
blk.10.layer_output_norm.weight
F32
F32
[768]
blk.11
blk.11.attn_k.bias
F32
F32
[768]
blk.11.attn_k.weight
F16
F16
[768, 768]
blk.11.attn_k_norm.bias
F32
F32
[768]
blk.11.attn_k_norm.weight
F32
F32
[768]
blk.11.attn_output.bias
F32
F32
[768]
blk.11.attn_output.weight
F16
F16
[768, 768]
blk.11.attn_output_norm.bias
F32
F32
[768]
blk.11.attn_output_norm.weight
F32
F32
[768]
blk.11.attn_q.bias
F32
F32
[768]
blk.11.attn_q.weight
F16
F16
[768, 768]
blk.11.attn_q_norm.bias
F32
F32
[768]
blk.11.attn_q_norm.weight
F32
F32
[768]
blk.11.attn_v.bias
F32
F32
[768]
blk.11.attn_v.weight
F16
F16
[768, 768]
blk.11.ffn_down.bias
F32
F32
[768]
blk.11.ffn_down.weight
F16
F16
[3072, 768]
blk.11.ffn_gate.weight
F16
F16
[768, 3072]
blk.11.ffn_up.weight
F16
F16
[768, 3072]
blk.11.layer_output_norm.bias
F32
F32
[768]
blk.11.layer_output_norm.weight
F32
F32
[768]
token_embd_norm.bias
F32
F32
[768]
token_embd_norm.weight
F32
F32
[768]
token_types.weight
F32
F32
[768, 2]