DBRX: A New State-of-the-Art Open LLM
Sharad Gupta
Linkedin Top Voice I Ex-McKinsey I GenAI Product and Growth leader in Banking, FinTech | Ex-CMO and Head of Data science Foodpanda (Unicorn) I Ex-CBO and Product leader Tookitaki
DBRX is a new open-source large language model (LLM) created by Databricks that sets a new state-of-the-art on several benchmarks. Across standard benchmarks like MMLU, HumanEval, and GSM8K, DBRX Instruct outperforms established open models like LLaMA2, Grok, and Mixtral.DBRX also surpasses or matches the performance of leading closed models like GPT-3.5, Gemini, and Mistral Medium on many tasks.
?Training Efficiency and Compute Savings
Performance on Long-Context and Retrieval Tasks
DBRX Instruct was trained with up to a 32K token context window. It compares its performance to Mixtral Instruct and the latest GPT-3.5 Turbo and GPT-4 Turbo APIs on long-context benchmarks. On retrieval-augmented generation (RAG) tasks using a Wikipedia corpus, DBRX Instruct is competitive with other open and closed models.
*Averages for GPT-3.5 Turbo include only contexts up to 6K, as it supports a maximum of 6K.
The information provided discusses the inference efficiency of DBRX and similar models using NVIDIA TensorRT-LLM with optimized serving infrastructure and 6-bit precision. The benchmark aims to simulate real-world usage, including multiple users hitting the same inference server. Each user request contains an approximately 2000 token prompt, and each response comprises 256 tokens.
Users leveraging Databricks Foundation Model APIs can anticipate up to 50 tokens per second for DBRX on the optimized model serving platform with 8-bit quantization.
In summary, DBRX Instruct performs competitively with leading models on long-context and retrieval-augmented tasks, demonstrating its strong capabilities across various benchmarks.
DBRX represents a significant advancement in open-source large language models, providing state-of-the-art performance across a range of benchmarks while also being more efficient to train and use than previous models.