
30 years a hacker
-
mixtral-8x22b
The Mixtral-8x22B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts.
422 Pulls 3 Tags Updated 1 year ago
-
r1-1776
R1 1776 is a DeepSeek-R1 reasoning model that has been post-trained by Perplexity AI to remove Chinese Communist Party censorship.
222 Pulls 2 Tags Updated 6 months ago
-
deepseek-r1-671b-1.58bit
Unsloth quantized 1.58bit version of deepseek r1
222 Pulls 1 Tag Updated 6 months ago
-
wizardlm-2-8x22
Original version released Apritl 15th of WizardLM 2
198 Pulls 2 Tags Updated 1 year ago
-
qra-13b
Qra is foundation language model trained with causal language modeling objective on a large corpus of texts.
124 Pulls 2 Tags Updated 1 year ago
-
zephyr-orpo-141b-a35b-v0.1
Zephyr is a series of language models that are trained to act as helpful assistants.
95 Pulls 3 Tags Updated 1 year ago