The IBM Granite 2B and 8B models are text-only dense LLMs trained on over 12 trillion tokens of data, demonstrated significant improvements over their predecessors in performance and speed in IBM’s initial testing.
3,079 Pulls Updated 3 days ago
Updated 3 days ago
3 days ago
990a904bdaf3 · 3.6GB
Readme
Granite dense models
The IBM Granite 2B and 8B models are text-only dense LLMs trained on over 12 trillion tokens of data, demonstrated significant improvements over their predecessors in performance and speed in IBM’s initial testing.
They are designed to support tool-based use cases and for retrieval augmented generation (RAG), streamlining code generation, translation and bug fixing.
Parameter Sizes
2B:
ollama run granite3.1-dense:2b
8B:
ollama run granite3.1-dense:8b
Supported Languages
English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, Chinese (Simplified)
Capabilities
- Summarization
- Text classification
- Text extraction
- Question-answering
- Retrieval Augmented Generation (RAG)
- Code related tasks
- Function-calling tasks
- Multilingual dialog use cases
- Long-context tasks including long document/meeting summarization, long document QA, etc.
Granite mixture of experts models
The Granite mixture of experts models are available in 1B and 3B parameter sizes designed for low latency usage.
Learn more
- Developers: IBM Research
- GitHub Repository: ibm-granite/granite-language-models
- Website: Granite Docs
- Release Date: December 18th, 2024
- License: Apache 2.0.