TCYZ/ cokertme:6.7m

71 2 months ago

Türkiye'nin yapay zeka hamlesinde stratejik bir boşluğu dolduran Çökertme serisi, devasa modellerin aksine "her cihazda çalışan zeka" mottosuyla geliştirilmiştir. TCYZ projesi kapsamında sunulan bu aile, en küçük donanımla çalışabilir.

1m 1.6m 6.7m 28m 57m
32e2fbf8837e · 28MB
    Metadata
  • general.architecture
    llama
  • llama.attention.head_count
    8
  • llama.attention.head_count_kv
    8
  • llama.attention.layer_norm_rms_epsilon
    1e-05
  • llama.block_count
    6
  • llama.context_length
    128
  • llama.embedding_length
    64
  • llama.feed_forward_length
    170
  • tokenizer.ggml.model
    llama
  • tokenizer.ggml.scores
    [0, 0, 0, 0, 0, ...]
  • tokenizer.ggml.token_type
    [1, 1, 1, 1, 1, ...]
  • tokenizer.ggml.tokens
    [!, ", #, $, %, ...]
  • Tensor
  • token_embd.weight
    F32
    [64, 50257]
  • blk.0
  • blk.0.attn_k.weight
    F32
    [64, 64]
  • blk.0.attn_norm.weight
    F32
    [64]
  • blk.0.attn_output.weight
    F32
    [64, 64]
  • blk.0.attn_q.weight
    F32
    [64, 64]
  • blk.0.attn_v.weight
    F32
    [64, 64]
  • blk.0.ffn_down.weight
    F32
    [170, 64]
  • blk.0.ffn_gate.weight
    F32
    [64, 170]
  • blk.0.ffn_norm.weight
    F32
    [64]
  • blk.0.ffn_up.weight
    F32
    [64, 170]
  • blk.1
  • blk.1.attn_k.weight
    F32
    [64, 64]
  • blk.1.attn_norm.weight
    F32
    [64]
  • blk.1.attn_output.weight
    F32
    [64, 64]
  • blk.1.attn_q.weight
    F32
    [64, 64]
  • blk.1.attn_v.weight
    F32
    [64, 64]
  • blk.1.ffn_down.weight
    F32
    [170, 64]
  • blk.1.ffn_gate.weight
    F32
    [64, 170]
  • blk.1.ffn_norm.weight
    F32
    [64]
  • blk.1.ffn_up.weight
    F32
    [64, 170]
  • blk.2
  • blk.2.attn_k.weight
    F32
    [64, 64]
  • blk.2.attn_norm.weight
    F32
    [64]
  • blk.2.attn_output.weight
    F32
    [64, 64]
  • blk.2.attn_q.weight
    F32
    [64, 64]
  • blk.2.attn_v.weight
    F32
    [64, 64]
  • blk.2.ffn_down.weight
    F32
    [170, 64]
  • blk.2.ffn_gate.weight
    F32
    [64, 170]
  • blk.2.ffn_norm.weight
    F32
    [64]
  • blk.2.ffn_up.weight
    F32
    [64, 170]
  • blk.3
  • blk.3.attn_k.weight
    F32
    [64, 64]
  • blk.3.attn_norm.weight
    F32
    [64]
  • blk.3.attn_output.weight
    F32
    [64, 64]
  • blk.3.attn_q.weight
    F32
    [64, 64]
  • blk.3.attn_v.weight
    F32
    [64, 64]
  • blk.3.ffn_down.weight
    F32
    [170, 64]
  • blk.3.ffn_gate.weight
    F32
    [64, 170]
  • blk.3.ffn_norm.weight
    F32
    [64]
  • blk.3.ffn_up.weight
    F32
    [64, 170]
  • blk.4
  • blk.4.attn_k.weight
    F32
    [64, 64]
  • blk.4.attn_norm.weight
    F32
    [64]
  • blk.4.attn_output.weight
    F32
    [64, 64]
  • blk.4.attn_q.weight
    F32
    [64, 64]
  • blk.4.attn_v.weight
    F32
    [64, 64]
  • blk.4.ffn_down.weight
    F32
    [170, 64]
  • blk.4.ffn_gate.weight
    F32
    [64, 170]
  • blk.4.ffn_norm.weight
    F32
    [64]
  • blk.4.ffn_up.weight
    F32
    [64, 170]
  • blk.5
  • blk.5.attn_k.weight
    F32
    [64, 64]
  • blk.5.attn_norm.weight
    F32
    [64]
  • blk.5.attn_output.weight
    F32
    [64, 64]
  • blk.5.attn_q.weight
    F32
    [64, 64]
  • blk.5.attn_v.weight
    F32
    [64, 64]
  • blk.5.ffn_down.weight
    F32
    [170, 64]
  • blk.5.ffn_gate.weight
    F32
    [64, 170]
  • blk.5.ffn_norm.weight
    F32
    [64]
  • blk.5.ffn_up.weight
    F32
    [64, 170]
  • output.weight
    F32
    [64, 50257]
  • output_norm.weight
    F32
    [64]