Databricks Releases DBRX, a State-of-the-Art Generative AI LLM, Under a Semi-Open Source License

GPT-3.5-beating model available to all, so long as you don't use it to train other models or support more than 700 million monthly users.

Data lake specialist Databricks has announced the release of a semi-open source large language model (LLM), DBRX, which it claims sets a "new standard" for generative artificial intelligence (gen AI) — and that, by the company's own testing, outperforms rivals including Llama2, Mixtral, Grok, and OpenAI's GPT-3.5.

"Databricks' mission is to deliver data intelligence to every enterprise by allowing organizations to understand and use their unique data to build their own AI systems," the company claims in its announcement of the new LLM. "Today, we are excited to advance our mission by open sourcing DBRX, a general purpose large language model (LLM) built by our Mosaic Research team that outperforms all established open source models on standard benchmarks. We believe that pushing the boundary of open source models enables generative AI for all enterprises that is customizable and transparent."

Based on a mixture-of-experts (MoE) model created using the company's open source MegaBlocks library, DBRX is claimed to offer improved performance by splitting itself into chunks depending on requirements — with the model itself being sized at an impressive 132 billion parameters, but only using 36 billion parameters at any given time to boost the throughput in tokens per second.

Despite this, Databricks claims the model outperforms its competition at a range of tasks — using, admittedly, its own Gauntlet benchmark suite. Testing on language understanding, programming, and math tasks, DBRX is claimed to beat rival open source models Llama2-70B, Mixtral, and Grok-1, as well as OpenAI's GPT-3.5 — nearly doubling the latter's score for programming tasks.

"[We] believe that open source LLMs will continue gaining momentum," Databricks claims in support of its launch. "In particular, we think they provide an exciting opportunity for organizations to customize open source LLMs that can become their IP, which they use to be competitive in their industry."

DBRX has been released under the custom Databricks Open Model License, which allows for reproduction and distribution but which specifically excludes using DBRX, derivatives, or outputs of same "to improve any other large language model" — and which includes a limit of 700 million monthly active users, after which a license must be requested at unspecified cost.

The company also requires DBRX users to agree to an acceptable use policy, which includes a moratorium on, among other things, using the model to provide medical advice "that is intended to be a substitute for professional medical advice, diagnosis, or treatment" or to "generate or disseminate information and place the information in any public context without expressly and intelligibly disclaiming that the information and/or content is machine generated."

If the restrictive covenants of the "open" license aren't a deal-breaker, DBRX is available on GitHub and Hugging Face now; more information on the model is available in Databricks' technical blog post.

Gareth Halfacree
Freelance journalist, technical author, hacker, tinkerer, erstwhile sysadmin. For hire: freelance@halfacree.co.uk.
Latest articles
Sponsored articles
Related articles
Latest articles
Read more
Related articles