Ollama
Models Docs Pricing
Sign in Download
Models Download Docs Pricing Sign in
⇅
Ollama
Search for models on Ollama.
  • mistral-nemo

    A state-of-the-art 12B model with 128k context length, built by Mistral AI in collaboration with NVIDIA.

    tools 12b

    3.9M  Pulls 17  Tags Updated  8 months ago

  • falcon3

    A family of efficient AI models under 10B parameters performant in science, math, and coding through innovative training techniques.

    1b 3b 7b 10b

    2.5M  Pulls 17  Tags Updated  1 year ago

  • llama2

    Llama 2 is a collection of foundation language models ranging from 7B to 70B parameters.

    7b 13b 70b

    6.5M  Pulls 102  Tags Updated  2 years ago

  • llama4

    Meta's latest collection of multimodal models.

    vision tools 16x17b 128x17b

    1.6M  Pulls 11  Tags Updated  10 months ago

  • mistral-small

    Mistral Small 3 sets a new benchmark in the “small” Large Language Models category below 70B.

    tools 22b 24b

    2.9M  Pulls 21  Tags Updated  1 year ago

  • qwq

    QwQ is the reasoning model of the Qwen series.

    tools 32b

    2.2M  Pulls 8  Tags Updated  1 year ago

  • tinyllama

    The TinyLlama project is an open endeavor to train a compact 1.1B Llama model on 3 trillion tokens.

    1.1b

    4.6M  Pulls 36  Tags Updated  2 years ago

  • codellama

    A large language model that can use text prompts to generate and discuss code.

    7b 13b 34b 70b

    5.2M  Pulls 199  Tags Updated  1 year ago

  • deepseek-coder

    DeepSeek Coder is a capable coding model trained on two trillion code and natural language tokens.

    1.3b 6.7b 33b

    3.9M  Pulls 102  Tags Updated  2 years ago

  • snowflake-arctic-embed

    A suite of text embedding models by Snowflake, optimized for performance.

    embedding 22m 33m 110m 137m 335m

    2.9M  Pulls 16  Tags Updated  1 year ago

  • codegemma

    CodeGemma is a collection of powerful, lightweight models that can perform a variety of coding tasks like fill-in-the-middle code completion, code generation, natural language understanding, mathematical reasoning, and instruction following.

    2b 7b

    2.9M  Pulls 85  Tags Updated  1 year ago

  • deepscaler

    A fine-tuned version of Deepseek-R1-Distilled-Qwen-1.5B that surpasses the performance of OpenAI’s o1-preview with just 1.5B parameters on popular math evaluations.

    1.5b

    1.2M  Pulls 5  Tags Updated  1 year ago

  • deepseek-coder-v2

    An open-source Mixture-of-Experts code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks.

    16b 236b

    2.3M  Pulls 64  Tags Updated  1 year ago

  • all-minilm

    Embedding models on very large sentence level datasets.

    embedding 22m 33m

    2.9M  Pulls 10  Tags Updated  1 year ago

  • starcoder2

    StarCoder2 is the next generation of transparently trained open code LLMs that comes in three sizes: 3B, 7B and 15B parameters.

    3b 7b 15b

    2.7M  Pulls 67  Tags Updated  1 year ago

  • phi4-mini

    Phi-4-mini brings significant enhancements in multilingual support, reasoning, and mathematics, and now, the long-awaited function calling feature is finally supported.

    tools 3.8b

    1.1M  Pulls 5  Tags Updated  1 year ago

  • llava-llama3

    A LLaVA model fine-tuned from Llama 3 Instruct with better scores in several benchmarks.

    vision 8b

    2.2M  Pulls 4  Tags Updated  1 year ago

  • openthinker

    A fully open-source family of reasoning models built using a dataset derived by distilling DeepSeek-R1.

    7b 32b

    1.1M  Pulls 15  Tags Updated  1 year ago

  • deepcoder

    DeepCoder is a fully open-Source 14B coder model at O3-mini level, with a 1.5B version also available.

    1.5b 14b

    841.6K  Pulls 9  Tags Updated  1 year ago

  • smollm

    🪐 A family of small models with 135M, 360M, and 1.7B parameters, trained on a new high-quality dataset.

    135m 360m 1.7b

    1.6M  Pulls 94  Tags Updated  1 year ago

© 2026 Ollama
Blog Contact