BoredFjord company
-
gigachat3-10B-A1.8
GigaChat3-10B-A1.8B is a dialogue model of the GigaChat family. The model is based on a Mixture-of-Experts (MoE) architecture with 10B total and 1.8B active parameters. The architecture includes Multi-head Latent Attention and Multi-Token Prediction.
104 Pulls 1 Tag Updated 1 month ago
-
GigaChat3.1-10B-A1.8B-q4_K_M
GigaChat 3.1 Lightning is the compact instruct model of the GigaChat 3.1 family. It is a Mixture-of-Experts (MoE) model with 10B total parameters and 1.8B active parameters, designed for fast multilingual assistant workloads, reasoning, code
30 Pulls 1 Tag Updated yesterday