A general-purpose multimodal mixture-of-experts model for production-grade tasks and enterprise workloads.
41.1K Pulls 1 Tag Updated 4 months ago
An update to Mistral Small that improves on function calling, instruction following, and less repetition errors.
1.7M Pulls 5 Tags Updated 9 months ago
The 7B model released by Mistral AI, updated to version 0.3.
28.1M Pulls 84 Tags Updated 9 months ago
A state-of-the-art 12B model with 128k context length, built by Mistral AI in collaboration with NVIDIA.
3.8M Pulls 17 Tags Updated 8 months ago
Mistral Small 3 sets a new benchmark in the “small” Large Language Models category below 70B.
2.8M Pulls 21 Tags Updated 1 year ago
Building upon Mistral Small 3, Mistral Small 3.1 (2503) adds state-of-the-art vision understanding and enhances long context capabilities up to 128k tokens without compromising text performance.
707.1K Pulls 5 Tags Updated 1 year ago
Mistral Large 2 is Mistral's new flagship model that is significantly more capable in code generation, mathematics, and reasoning with 128k context window and support for dozens of languages.
1.1M Pulls 32 Tags Updated 1 year ago
Mistral OpenOrca is a 7 billion parameter model, fine-tuned on top of the Mistral 7B model using the OpenOrca dataset.
607.1K Pulls 17 Tags Updated 2 years ago
MistralLite is a fine-tuned model based on Mistral with enhanced capabilities of processing long contexts.
458.5K Pulls 17 Tags Updated 2 years ago
A set of Mixture of Experts (MoE) model with open weights by Mistral AI in 8x7b and 8x22b parameter sizes.
2.5M Pulls 70 Tags Updated 1 year ago
Codestral is Mistral AI’s first-ever code model designed for code generation tasks.
1.1M Pulls 17 Tags Updated 1 year ago
The uncensored Dolphin model based on Mistral that excels at coding tasks. Updated to version 2.8.
1.3M Pulls 120 Tags Updated 2 years ago
OpenHermes 2.5 is a 7B model fine-tuned by Teknium on Mistral with fully open datasets.
935.7K Pulls 35 Tags Updated 2 years ago
Zephyr is a series of fine-tuned versions of the Mistral and Mixtral models that are trained to act as helpful assistants.
1.1M Pulls 40 Tags Updated 1 year ago
A fine-tuned model based on Mistral with good coverage of domain and language.
901.9K Pulls 50 Tags Updated 2 years ago
BakLLaVA is a multimodal model consisting of the Mistral 7B base model augmented with the LLaVA architecture.
793.1K Pulls 17 Tags Updated 2 years ago
A companion assistant trained in philosophy, psychology, and personal relationships. Based on Mistral.
844.4K Pulls 49 Tags Updated 2 years ago
An extension of Mistral to support context windows of 64K or 128K.
772.7K Pulls 33 Tags Updated 2 years ago
MathΣtral: a 7B model designed for math reasoning and scientific discovery by Mistral AI.
476.5K Pulls 17 Tags Updated 1 year ago
The Ministral 3 family is designed for edge deployment, capable of running on a wide range of hardware.
899.2K Pulls 16 Tags Updated 3 months ago