A general-purpose multimodal mixture-of-experts model for production-grade tasks and enterprise workloads.
18.9K Pulls 1 Tag Updated 2 months ago
An update to Mistral Small that improves on function calling, instruction following, and less repetition errors.
1.3M Pulls 5 Tags Updated 8 months ago
Building upon Mistral Small 3, Mistral Small 3.1 (2503) adds state-of-the-art vision understanding and enhances long context capabilities up to 128k tokens without compromising text performance.
603.1K Pulls 5 Tags Updated 10 months ago
The 7B model released by Mistral AI, updated to version 0.3.
25.5M Pulls 84 Tags Updated 7 months ago
A state-of-the-art 12B model with 128k context length, built by Mistral AI in collaboration with NVIDIA.
3.4M Pulls 17 Tags Updated 7 months ago
Mistral Small 3 sets a new benchmark in the “small” Large Language Models category below 70B.
2.4M Pulls 21 Tags Updated 1 year ago
Mistral Large 2 is Mistral's new flagship model that is significantly more capable in code generation, mathematics, and reasoning with 128k context window and support for dozens of languages.
549.2K Pulls 32 Tags Updated 1 year ago
Mistral OpenOrca is a 7 billion parameter model, fine-tuned on top of the Mistral 7B model using the OpenOrca dataset.
349.1K Pulls 17 Tags Updated 2 years ago
MistralLite is a fine-tuned model based on Mistral with enhanced capabilities of processing long contexts.
205.2K Pulls 17 Tags Updated 2 years ago
A set of Mixture of Experts (MoE) model with open weights by Mistral AI in 8x7b and 8x22b parameter sizes.
1.9M Pulls 70 Tags Updated 1 year ago
Codestral is Mistral AI’s first-ever code model designed for code generation tasks.
825.2K Pulls 17 Tags Updated 1 year ago
The uncensored Dolphin model based on Mistral that excels at coding tasks. Updated to version 2.8.
736.1K Pulls 120 Tags Updated 1 year ago
BakLLaVA is a multimodal model consisting of the Mistral 7B base model augmented with the LLaVA architecture.
518.7K Pulls 17 Tags Updated 2 years ago
Zephyr is a series of fine-tuned versions of the Mistral and Mixtral models that are trained to act as helpful assistants.
559.2K Pulls 40 Tags Updated 1 year ago
OpenHermes 2.5 is a 7B model fine-tuned by Teknium on Mistral with fully open datasets.
454.2K Pulls 35 Tags Updated 2 years ago
A fine-tuned model based on Mistral with good coverage of domain and language.
422.9K Pulls 50 Tags Updated 2 years ago
A companion assistant trained in philosophy, psychology, and personal relationships. Based on Mistral.
379.1K Pulls 49 Tags Updated 2 years ago
MathΣtral: a 7B model designed for math reasoning and scientific discovery by Mistral AI.
219.2K Pulls 17 Tags Updated 1 year ago
An extension of Mistral to support context windows of 64K or 128K.
318.3K Pulls 33 Tags Updated 2 years ago
The Ministral 3 family is designed for edge deployment, capable of running on a wide range of hardware.
470.4K Pulls 16 Tags Updated 2 months ago