BGE-M3 is a new model from BAAI distinguished for its versatility in Multi-Functionality, Multi-Linguality, and Multi-Granularity.
4.1M Pulls 3 Tags Updated 1 year ago
Embedding model from BAAI mapping texts to vectors.
261.5K Pulls 3 Tags Updated 1 year ago
MedGemma 1.5 4B is an updated version of the MedGemma 4B model.
12.6K Pulls 5 Tags Updated 2 weeks ago
A new collection of open translation models built on Gemma 3, helping people communicate across 55 languages.
1.3M Pulls 13 Tags Updated 3 months ago
Dolphin 3.0 Llama 3.1 8B 🐬 is the next generation of the Dolphin series of instruct-tuned models designed to be the ultimate general purpose local model, enabling coding, math, agentic, function calling, and general use cases.
3.8M Pulls 5 Tags Updated 1 year ago
The IBM Granite 2B and 8B models are designed to support tool-based use cases and support for retrieval augmented generation (RAG), streamlining code generation, translation and bug fixing.
951.6K Pulls 33 Tags Updated 1 year ago
A model from NVIDIA based on Llama 3 that excels at conversational question answering (QA) and retrieval-augmented generation (RAG).
951.6K Pulls 35 Tags Updated 1 year ago
Llama-3.1-Nemotron-70B-Instruct is a large language model customized by NVIDIA to improve the helpfulness of LLM generated responses to user queries.
572.1K Pulls 17 Tags Updated 1 year ago
Stable LM 2 is a state-of-the-art 1.6B and 12B parameter language model trained on multilingual data in English, Spanish, German, Italian, French, Portuguese, and Dutch.
955.4K Pulls 84 Tags Updated 1 year ago
EXAONE 3.5 is a collection of instruction-tuned bilingual (English and Korean) generative models ranging from 2.4B to 32B parameters, developed and released by LG AI Research.
496.4K Pulls 13 Tags Updated 1 year ago
A large language model built by the Technology Innovation Institute (TII) for use in summarization, text generation, and chat bots.
1.1M Pulls 38 Tags Updated 2 years ago
An expansion of Llama 2 that specializes in integrating both general language understanding and domain-specific knowledge, particularly in programming and mathematics.
853K Pulls 33 Tags Updated 2 years ago
Merge of the Open Orca OpenChat model and the Garage-bAInd Platypus 2 model. Designed for chat and code generation.
480.6K Pulls 17 Tags Updated 2 years ago
Gemma 4 E2B fine-tuned for microscopy ID — rich 4-section scientific descriptions (genus + morphology + habitat + ID cues). Same 3.4 GB as v2, runs offline on laptop / phone. Apache 2.0. Research model — not a medical device.
10 Pulls 1 Tag Updated 17 hours ago
Gemma 4 E2B fine-tuned on 122k microscopy VQA · 145+ genera · 5 categories · runs offline on a sub-$100 phone · Unsloth + llama.cpp · Apache 2.0 · research/educational only, not a medical device
4 Pulls 1 Tag Updated 2 days ago
Gemma 4 26B MoE quantized by BatiAI. 77 t/s on M4 Max. Requires 24GB+ Mac.
2,079 Pulls 6 Tags Updated 3 weeks ago
1,889 Pulls 2 Tags Updated 3 weeks ago
1,660 Pulls 2 Tags Updated 3 weeks ago
600 Pulls 4 Tags Updated 3 weeks ago
As the strongest model in the 30B class, GLM-4.7-Flash offers a new option for lightweight deployment that balances performance and efficiency.
1.2M Pulls 4 Tags Updated 3 months ago