TCYZ/ cokertme:57m

71 2 months ago

Türkiye'nin yapay zeka hamlesinde stratejik bir boşluğu dolduran Çökertme serisi, devasa modellerin aksine "her cihazda çalışan zeka" mottosuyla geliştirilmiştir. TCYZ projesi kapsamında sunulan bu aile, en küçük donanımla çalışabilir.

1m 1.6m 6.7m 28m 57m
be2202913bec · 232MB
    Metadata
  • general.architecture
    llama
  • llama.attention.head_count
    8
  • llama.attention.head_count_kv
    8
  • llama.attention.layer_norm_rms_epsilon
    1e-05
  • llama.block_count
    2
  • llama.context_length
    32
  • llama.embedding_length
    512
  • llama.feed_forward_length
    1365
  • tokenizer.ggml.model
    llama
  • tokenizer.ggml.scores
    [0, 0, 0, 0, 0, ...]
  • tokenizer.ggml.token_type
    [1, 1, 1, 1, 1, ...]
  • tokenizer.ggml.tokens
    [!, ", #, $, %, ...]
  • Tensor
  • token_embd.weight
    F32
    [512, 50257]
  • blk.0
  • blk.0.attn_k.weight
    F32
    [512, 512]
  • blk.0.attn_norm.weight
    F32
    [512]
  • blk.0.attn_output.weight
    F32
    [512, 512]
  • blk.0.attn_q.weight
    F32
    [512, 512]
  • blk.0.attn_v.weight
    F32
    [512, 512]
  • blk.0.ffn_down.weight
    F32
    [1365, 512]
  • blk.0.ffn_gate.weight
    F32
    [512, 1365]
  • blk.0.ffn_norm.weight
    F32
    [512]
  • blk.0.ffn_up.weight
    F32
    [512, 1365]
  • blk.1
  • blk.1.attn_k.weight
    F32
    [512, 512]
  • blk.1.attn_norm.weight
    F32
    [512]
  • blk.1.attn_output.weight
    F32
    [512, 512]
  • blk.1.attn_q.weight
    F32
    [512, 512]
  • blk.1.attn_v.weight
    F32
    [512, 512]
  • blk.1.ffn_down.weight
    F32
    [1365, 512]
  • blk.1.ffn_gate.weight
    F32
    [512, 1365]
  • blk.1.ffn_norm.weight
    F32
    [512]
  • blk.1.ffn_up.weight
    F32
    [512, 1365]
  • output.weight
    F32
    [512, 50257]
  • output_norm.weight
    F32
    [512]