Models
GitHub
Discord
Docs
Pricing
Sign in
Download
Models
Download
GitHub
Discord
Docs
Pricing
Sign in
alibayram
/
embeddingmagibu-200m
:latest
6
Downloads
Updated
2 weeks ago
Bu model, Türkçe odaklı çok uzun bağlam (8192 token) sentence embedding üretmek için eğitilmiş bir SentenceTransformer modelidir. 768 boyutlu normalize vektör uzayına projeksiyon yapar.
Bu model, Türkçe odaklı çok uzun bağlam (8192 token) sentence embedding üretmek için eğitilmiş bir SentenceTransformer modelidir. 768 boyutlu normalize vektör uzayına projeksiyon yapar.
Cancel
embedding
embeddingmagibu-200m:latest
...
/
model
1c1fc463bc12 · 411MB
Metadata
general.architecture
gemma-embedding
gemma-embedding
general.file_type
BF16
BF16
gemma-embedding.attention.head_count
3
3
gemma-embedding.attention.head_count_kv
1
1
gemma-embedding.attention.key_length
256
256
gemma-embedding.attention.layer_norm_rms_epsilon
1e-06
1e-06
gemma-embedding.attention.sliding_window
3
3
gemma-embedding.attention.value_length
256
256
gemma-embedding.block_count
24
24
gemma-embedding.context_length
8192
8192
gemma-embedding.embedding_length
768
768
gemma-embedding.feed_forward_length
1152
1152
gemma-embedding.pooling_type
Mean
Mean
gemma-embedding.rope.freq_base
1e+06
1e+06
gemma-embedding.rope.freq_base_swa
10000
10000
tokenizer.ggml.add_bos_token
true
true
tokenizer.ggml.add_sep_token
false
false
tokenizer.ggml.bos_token_id
2
2
tokenizer.ggml.eos_token_id
1
1
tokenizer.ggml.mask_token_id
4
4
tokenizer.ggml.merges
[ , ▁ ▁, , e r, i n, ...]
[ , ▁ ▁, , e r, i n, ...]
tokenizer.ggml.model
gpt2
gpt2
tokenizer.ggml.padding_token_id
0
0
tokenizer.ggml.pre
default
default
tokenizer.ggml.token_type
[3, 3, 3, 3, 3, ...]
[3, 3, 3, 3, 3, ...]
tokenizer.ggml.tokens
[<pad>, <eos>, <bos>, <unk>, <mask>, ...]
[<pad>, <eos>, <bos>, <unk>, <mask>, ...]
tokenizer.ggml.unknown_token_id
3
3
Tensor
Name
Type
Shape
token_embd.weight
BF16
BF16
[768, 131073]
blk.0
blk.0.attn_k.weight
BF16
BF16
[768, 256]
blk.0.attn_k_norm.weight
F32
F32
[256]
blk.0.attn_norm.weight
F32
F32
[768]
blk.0.attn_output.weight
BF16
BF16
[768, 768]
blk.0.attn_q.weight
BF16
BF16
[768, 768]
blk.0.attn_q_norm.weight
F32
F32
[256]
blk.0.attn_v.weight
BF16
BF16
[768, 256]
blk.0.ffn_down.weight
BF16
BF16
[1152, 768]
blk.0.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.0.ffn_norm.weight
F32
F32
[768]
blk.0.ffn_up.weight
BF16
BF16
[768, 1152]
blk.0.post_attention_norm.weight
F32
F32
[768]
blk.0.post_ffw_norm.weight
F32
F32
[768]
blk.1
blk.1.attn_k.weight
BF16
BF16
[768, 256]
blk.1.attn_k_norm.weight
F32
F32
[256]
blk.1.attn_norm.weight
F32
F32
[768]
blk.1.attn_output.weight
BF16
BF16
[768, 768]
blk.1.attn_q.weight
BF16
BF16
[768, 768]
blk.1.attn_q_norm.weight
F32
F32
[256]
blk.1.attn_v.weight
BF16
BF16
[768, 256]
blk.1.ffn_down.weight
BF16
BF16
[1152, 768]
blk.1.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.1.ffn_norm.weight
F32
F32
[768]
blk.1.ffn_up.weight
BF16
BF16
[768, 1152]
blk.1.post_attention_norm.weight
F32
F32
[768]
blk.1.post_ffw_norm.weight
F32
F32
[768]
blk.2
blk.2.attn_k.weight
BF16
BF16
[768, 256]
blk.2.attn_k_norm.weight
F32
F32
[256]
blk.2.attn_norm.weight
F32
F32
[768]
blk.2.attn_output.weight
BF16
BF16
[768, 768]
blk.2.attn_q.weight
BF16
BF16
[768, 768]
blk.2.attn_q_norm.weight
F32
F32
[256]
blk.2.attn_v.weight
BF16
BF16
[768, 256]
blk.2.ffn_down.weight
BF16
BF16
[1152, 768]
blk.2.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.2.ffn_norm.weight
F32
F32
[768]
blk.2.ffn_up.weight
BF16
BF16
[768, 1152]
blk.2.post_attention_norm.weight
F32
F32
[768]
blk.2.post_ffw_norm.weight
F32
F32
[768]
blk.3
blk.3.attn_k.weight
BF16
BF16
[768, 256]
blk.3.attn_k_norm.weight
F32
F32
[256]
blk.3.attn_norm.weight
F32
F32
[768]
blk.3.attn_output.weight
BF16
BF16
[768, 768]
blk.3.attn_q.weight
BF16
BF16
[768, 768]
blk.3.attn_q_norm.weight
F32
F32
[256]
blk.3.attn_v.weight
BF16
BF16
[768, 256]
blk.3.ffn_down.weight
BF16
BF16
[1152, 768]
blk.3.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.3.ffn_norm.weight
F32
F32
[768]
blk.3.ffn_up.weight
BF16
BF16
[768, 1152]
blk.3.post_attention_norm.weight
F32
F32
[768]
blk.3.post_ffw_norm.weight
F32
F32
[768]
blk.4
blk.4.attn_k.weight
BF16
BF16
[768, 256]
blk.4.attn_k_norm.weight
F32
F32
[256]
blk.4.attn_norm.weight
F32
F32
[768]
blk.4.attn_output.weight
BF16
BF16
[768, 768]
blk.4.attn_q.weight
BF16
BF16
[768, 768]
blk.4.attn_q_norm.weight
F32
F32
[256]
blk.4.attn_v.weight
BF16
BF16
[768, 256]
blk.4.ffn_down.weight
BF16
BF16
[1152, 768]
blk.4.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.4.ffn_norm.weight
F32
F32
[768]
blk.4.ffn_up.weight
BF16
BF16
[768, 1152]
blk.4.post_attention_norm.weight
F32
F32
[768]
blk.4.post_ffw_norm.weight
F32
F32
[768]
blk.5
blk.5.attn_k.weight
BF16
BF16
[768, 256]
blk.5.attn_k_norm.weight
F32
F32
[256]
blk.5.attn_norm.weight
F32
F32
[768]
blk.5.attn_output.weight
BF16
BF16
[768, 768]
blk.5.attn_q.weight
BF16
BF16
[768, 768]
blk.5.attn_q_norm.weight
F32
F32
[256]
blk.5.attn_v.weight
BF16
BF16
[768, 256]
blk.5.ffn_down.weight
BF16
BF16
[1152, 768]
blk.5.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.5.ffn_norm.weight
F32
F32
[768]
blk.5.ffn_up.weight
BF16
BF16
[768, 1152]
blk.5.post_attention_norm.weight
F32
F32
[768]
blk.5.post_ffw_norm.weight
F32
F32
[768]
blk.6
blk.6.attn_k.weight
BF16
BF16
[768, 256]
blk.6.attn_k_norm.weight
F32
F32
[256]
blk.6.attn_norm.weight
F32
F32
[768]
blk.6.attn_output.weight
BF16
BF16
[768, 768]
blk.6.attn_q.weight
BF16
BF16
[768, 768]
blk.6.attn_q_norm.weight
F32
F32
[256]
blk.6.attn_v.weight
BF16
BF16
[768, 256]
blk.6.ffn_down.weight
BF16
BF16
[1152, 768]
blk.6.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.6.ffn_norm.weight
F32
F32
[768]
blk.6.ffn_up.weight
BF16
BF16
[768, 1152]
blk.6.post_attention_norm.weight
F32
F32
[768]
blk.6.post_ffw_norm.weight
F32
F32
[768]
blk.7
blk.7.attn_k.weight
BF16
BF16
[768, 256]
blk.7.attn_k_norm.weight
F32
F32
[256]
blk.7.attn_norm.weight
F32
F32
[768]
blk.7.attn_output.weight
BF16
BF16
[768, 768]
blk.7.attn_q.weight
BF16
BF16
[768, 768]
blk.7.attn_q_norm.weight
F32
F32
[256]
blk.7.attn_v.weight
BF16
BF16
[768, 256]
blk.7.ffn_down.weight
BF16
BF16
[1152, 768]
blk.7.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.7.ffn_norm.weight
F32
F32
[768]
blk.7.ffn_up.weight
BF16
BF16
[768, 1152]
blk.7.post_attention_norm.weight
F32
F32
[768]
blk.7.post_ffw_norm.weight
F32
F32
[768]
blk.8
blk.8.attn_k.weight
BF16
BF16
[768, 256]
blk.8.attn_k_norm.weight
F32
F32
[256]
blk.8.attn_norm.weight
F32
F32
[768]
blk.8.attn_output.weight
BF16
BF16
[768, 768]
blk.8.attn_q.weight
BF16
BF16
[768, 768]
blk.8.attn_q_norm.weight
F32
F32
[256]
blk.8.attn_v.weight
BF16
BF16
[768, 256]
blk.8.ffn_down.weight
BF16
BF16
[1152, 768]
blk.8.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.8.ffn_norm.weight
F32
F32
[768]
blk.8.ffn_up.weight
BF16
BF16
[768, 1152]
blk.8.post_attention_norm.weight
F32
F32
[768]
blk.8.post_ffw_norm.weight
F32
F32
[768]
blk.9
blk.9.attn_k.weight
BF16
BF16
[768, 256]
blk.9.attn_k_norm.weight
F32
F32
[256]
blk.9.attn_norm.weight
F32
F32
[768]
blk.9.attn_output.weight
BF16
BF16
[768, 768]
blk.9.attn_q.weight
BF16
BF16
[768, 768]
blk.9.attn_q_norm.weight
F32
F32
[256]
blk.9.attn_v.weight
BF16
BF16
[768, 256]
blk.9.ffn_down.weight
BF16
BF16
[1152, 768]
blk.9.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.9.ffn_norm.weight
F32
F32
[768]
blk.9.ffn_up.weight
BF16
BF16
[768, 1152]
blk.9.post_attention_norm.weight
F32
F32
[768]
blk.9.post_ffw_norm.weight
F32
F32
[768]
blk.10
blk.10.attn_k.weight
BF16
BF16
[768, 256]
blk.10.attn_k_norm.weight
F32
F32
[256]
blk.10.attn_norm.weight
F32
F32
[768]
blk.10.attn_output.weight
BF16
BF16
[768, 768]
blk.10.attn_q.weight
BF16
BF16
[768, 768]
blk.10.attn_q_norm.weight
F32
F32
[256]
blk.10.attn_v.weight
BF16
BF16
[768, 256]
blk.10.ffn_down.weight
BF16
BF16
[1152, 768]
blk.10.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.10.ffn_norm.weight
F32
F32
[768]
blk.10.ffn_up.weight
BF16
BF16
[768, 1152]
blk.10.post_attention_norm.weight
F32
F32
[768]
blk.10.post_ffw_norm.weight
F32
F32
[768]
blk.11
blk.11.attn_k.weight
BF16
BF16
[768, 256]
blk.11.attn_k_norm.weight
F32
F32
[256]
blk.11.attn_norm.weight
F32
F32
[768]
blk.11.attn_output.weight
BF16
BF16
[768, 768]
blk.11.attn_q.weight
BF16
BF16
[768, 768]
blk.11.attn_q_norm.weight
F32
F32
[256]
blk.11.attn_v.weight
BF16
BF16
[768, 256]
blk.11.ffn_down.weight
BF16
BF16
[1152, 768]
blk.11.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.11.ffn_norm.weight
F32
F32
[768]
blk.11.ffn_up.weight
BF16
BF16
[768, 1152]
blk.11.post_attention_norm.weight
F32
F32
[768]
blk.11.post_ffw_norm.weight
F32
F32
[768]
blk.12
blk.12.attn_k.weight
BF16
BF16
[768, 256]
blk.12.attn_k_norm.weight
F32
F32
[256]
blk.12.attn_norm.weight
F32
F32
[768]
blk.12.attn_output.weight
BF16
BF16
[768, 768]
blk.12.attn_q.weight
BF16
BF16
[768, 768]
blk.12.attn_q_norm.weight
F32
F32
[256]
blk.12.attn_v.weight
BF16
BF16
[768, 256]
blk.12.ffn_down.weight
BF16
BF16
[1152, 768]
blk.12.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.12.ffn_norm.weight
F32
F32
[768]
blk.12.ffn_up.weight
BF16
BF16
[768, 1152]
blk.12.post_attention_norm.weight
F32
F32
[768]
blk.12.post_ffw_norm.weight
F32
F32
[768]
blk.13
blk.13.attn_k.weight
BF16
BF16
[768, 256]
blk.13.attn_k_norm.weight
F32
F32
[256]
blk.13.attn_norm.weight
F32
F32
[768]
blk.13.attn_output.weight
BF16
BF16
[768, 768]
blk.13.attn_q.weight
BF16
BF16
[768, 768]
blk.13.attn_q_norm.weight
F32
F32
[256]
blk.13.attn_v.weight
BF16
BF16
[768, 256]
blk.13.ffn_down.weight
BF16
BF16
[1152, 768]
blk.13.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.13.ffn_norm.weight
F32
F32
[768]
blk.13.ffn_up.weight
BF16
BF16
[768, 1152]
blk.13.post_attention_norm.weight
F32
F32
[768]
blk.13.post_ffw_norm.weight
F32
F32
[768]
blk.14
blk.14.attn_k.weight
BF16
BF16
[768, 256]
blk.14.attn_k_norm.weight
F32
F32
[256]
blk.14.attn_norm.weight
F32
F32
[768]
blk.14.attn_output.weight
BF16
BF16
[768, 768]
blk.14.attn_q.weight
BF16
BF16
[768, 768]
blk.14.attn_q_norm.weight
F32
F32
[256]
blk.14.attn_v.weight
BF16
BF16
[768, 256]
blk.14.ffn_down.weight
BF16
BF16
[1152, 768]
blk.14.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.14.ffn_norm.weight
F32
F32
[768]
blk.14.ffn_up.weight
BF16
BF16
[768, 1152]
blk.14.post_attention_norm.weight
F32
F32
[768]
blk.14.post_ffw_norm.weight
F32
F32
[768]
blk.15
blk.15.attn_k.weight
BF16
BF16
[768, 256]
blk.15.attn_k_norm.weight
F32
F32
[256]
blk.15.attn_norm.weight
F32
F32
[768]
blk.15.attn_output.weight
BF16
BF16
[768, 768]
blk.15.attn_q.weight
BF16
BF16
[768, 768]
blk.15.attn_q_norm.weight
F32
F32
[256]
blk.15.attn_v.weight
BF16
BF16
[768, 256]
blk.15.ffn_down.weight
BF16
BF16
[1152, 768]
blk.15.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.15.ffn_norm.weight
F32
F32
[768]
blk.15.ffn_up.weight
BF16
BF16
[768, 1152]
blk.15.post_attention_norm.weight
F32
F32
[768]
blk.15.post_ffw_norm.weight
F32
F32
[768]
blk.16
blk.16.attn_k.weight
BF16
BF16
[768, 256]
blk.16.attn_k_norm.weight
F32
F32
[256]
blk.16.attn_norm.weight
F32
F32
[768]
blk.16.attn_output.weight
BF16
BF16
[768, 768]
blk.16.attn_q.weight
BF16
BF16
[768, 768]
blk.16.attn_q_norm.weight
F32
F32
[256]
blk.16.attn_v.weight
BF16
BF16
[768, 256]
blk.16.ffn_down.weight
BF16
BF16
[1152, 768]
blk.16.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.16.ffn_norm.weight
F32
F32
[768]
blk.16.ffn_up.weight
BF16
BF16
[768, 1152]
blk.16.post_attention_norm.weight
F32
F32
[768]
blk.16.post_ffw_norm.weight
F32
F32
[768]
blk.17
blk.17.attn_k.weight
BF16
BF16
[768, 256]
blk.17.attn_k_norm.weight
F32
F32
[256]
blk.17.attn_norm.weight
F32
F32
[768]
blk.17.attn_output.weight
BF16
BF16
[768, 768]
blk.17.attn_q.weight
BF16
BF16
[768, 768]
blk.17.attn_q_norm.weight
F32
F32
[256]
blk.17.attn_v.weight
BF16
BF16
[768, 256]
blk.17.ffn_down.weight
BF16
BF16
[1152, 768]
blk.17.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.17.ffn_norm.weight
F32
F32
[768]
blk.17.ffn_up.weight
BF16
BF16
[768, 1152]
blk.17.post_attention_norm.weight
F32
F32
[768]
blk.17.post_ffw_norm.weight
F32
F32
[768]
blk.18
blk.18.attn_k.weight
BF16
BF16
[768, 256]
blk.18.attn_k_norm.weight
F32
F32
[256]
blk.18.attn_norm.weight
F32
F32
[768]
blk.18.attn_output.weight
BF16
BF16
[768, 768]
blk.18.attn_q.weight
BF16
BF16
[768, 768]
blk.18.attn_q_norm.weight
F32
F32
[256]
blk.18.attn_v.weight
BF16
BF16
[768, 256]
blk.18.ffn_down.weight
BF16
BF16
[1152, 768]
blk.18.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.18.ffn_norm.weight
F32
F32
[768]
blk.18.ffn_up.weight
BF16
BF16
[768, 1152]
blk.18.post_attention_norm.weight
F32
F32
[768]
blk.18.post_ffw_norm.weight
F32
F32
[768]
blk.19
blk.19.attn_k.weight
BF16
BF16
[768, 256]
blk.19.attn_k_norm.weight
F32
F32
[256]
blk.19.attn_norm.weight
F32
F32
[768]
blk.19.attn_output.weight
BF16
BF16
[768, 768]
blk.19.attn_q.weight
BF16
BF16
[768, 768]
blk.19.attn_q_norm.weight
F32
F32
[256]
blk.19.attn_v.weight
BF16
BF16
[768, 256]
blk.19.ffn_down.weight
BF16
BF16
[1152, 768]
blk.19.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.19.ffn_norm.weight
F32
F32
[768]
blk.19.ffn_up.weight
BF16
BF16
[768, 1152]
blk.19.post_attention_norm.weight
F32
F32
[768]
blk.19.post_ffw_norm.weight
F32
F32
[768]
blk.20
blk.20.attn_k.weight
BF16
BF16
[768, 256]
blk.20.attn_k_norm.weight
F32
F32
[256]
blk.20.attn_norm.weight
F32
F32
[768]
blk.20.attn_output.weight
BF16
BF16
[768, 768]
blk.20.attn_q.weight
BF16
BF16
[768, 768]
blk.20.attn_q_norm.weight
F32
F32
[256]
blk.20.attn_v.weight
BF16
BF16
[768, 256]
blk.20.ffn_down.weight
BF16
BF16
[1152, 768]
blk.20.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.20.ffn_norm.weight
F32
F32
[768]
blk.20.ffn_up.weight
BF16
BF16
[768, 1152]
blk.20.post_attention_norm.weight
F32
F32
[768]
blk.20.post_ffw_norm.weight
F32
F32
[768]
blk.21
blk.21.attn_k.weight
BF16
BF16
[768, 256]
blk.21.attn_k_norm.weight
F32
F32
[256]
blk.21.attn_norm.weight
F32
F32
[768]
blk.21.attn_output.weight
BF16
BF16
[768, 768]
blk.21.attn_q.weight
BF16
BF16
[768, 768]
blk.21.attn_q_norm.weight
F32
F32
[256]
blk.21.attn_v.weight
BF16
BF16
[768, 256]
blk.21.ffn_down.weight
BF16
BF16
[1152, 768]
blk.21.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.21.ffn_norm.weight
F32
F32
[768]
blk.21.ffn_up.weight
BF16
BF16
[768, 1152]
blk.21.post_attention_norm.weight
F32
F32
[768]
blk.21.post_ffw_norm.weight
F32
F32
[768]
blk.22
blk.22.attn_k.weight
BF16
BF16
[768, 256]
blk.22.attn_k_norm.weight
F32
F32
[256]
blk.22.attn_norm.weight
F32
F32
[768]
blk.22.attn_output.weight
BF16
BF16
[768, 768]
blk.22.attn_q.weight
BF16
BF16
[768, 768]
blk.22.attn_q_norm.weight
F32
F32
[256]
blk.22.attn_v.weight
BF16
BF16
[768, 256]
blk.22.ffn_down.weight
BF16
BF16
[1152, 768]
blk.22.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.22.ffn_norm.weight
F32
F32
[768]
blk.22.ffn_up.weight
BF16
BF16
[768, 1152]
blk.22.post_attention_norm.weight
F32
F32
[768]
blk.22.post_ffw_norm.weight
F32
F32
[768]
blk.23
blk.23.attn_k.weight
BF16
BF16
[768, 256]
blk.23.attn_k_norm.weight
F32
F32
[256]
blk.23.attn_norm.weight
F32
F32
[768]
blk.23.attn_output.weight
BF16
BF16
[768, 768]
blk.23.attn_q.weight
BF16
BF16
[768, 768]
blk.23.attn_q_norm.weight
F32
F32
[256]
blk.23.attn_v.weight
BF16
BF16
[768, 256]
blk.23.ffn_down.weight
BF16
BF16
[1152, 768]
blk.23.ffn_gate.weight
BF16
BF16
[768, 1152]
blk.23.ffn_norm.weight
F32
F32
[768]
blk.23.ffn_up.weight
BF16
BF16
[768, 1152]
blk.23.post_attention_norm.weight
F32
F32
[768]
blk.23.post_ffw_norm.weight
F32
F32
[768]
output_norm.weight
F32
F32
[768]