Powerful open-source LLM with exceptional performance on a wide range of benchmarks.
DBRX Instruct is a powerful, open-source large language model (LLM) developed by Databricks. It utilizes a fine-grained mixture-of-experts (MoE) architecture with 132 billion total parameters, of which 36 billion are active for any given input.
DBRX Instruct is a general-purpose LLM designed for a wide range of natural language processing tasks, including text generation, question answering, code generation, and more. It is particularly well-suited for applications requiring strong programming and mathematical reasoning capabilities.
DBRX Instruct is a multilingual model, supporting a wide range of languages.
DBRX Instruct is a transformer-based, decoder-only LLM that was trained using next-token prediction. It utilizes a fine-grained MoE architecture, with 16 experts and the ability to select 4 for any given input.
The model was pre-trained on 12 trillion tokens of carefully curated text and code data, with a maximum context length of 32,000 tokens. This dataset is estimated to be at least 2x better quality than the data used to train the MPT family of models.
On standard benchmarks, DBRX Instruct outperforms leading open models:
DBRX Instruct is available for use under the Databricks Open Model License.