DBRX is an open, general-purpose LLM created by Databricks.

132b

16.8K 7 months ago

Readme

DBRX requires Ollama 0.1.32

DBRX is a transformer-based decoder-only large language model (LLM) that was trained using next-token prediction. It uses a fine-grained mixture-of-experts (MoE) architecture with 132B total parameters of which 36B parameters are active on any input. It was pre-trained on 12T tokens of text and code data.

It is an especially capable code model, surpassing specialized models like CodeLLaMA-70B on programming, in addition to its strength as a general-purpose LLM.

References

Blog Post

HuggingFace