D

Dbrx Base

Developed by databricks
A Mixture of Experts (MoE) large language model developed by Databricks, with 132 billion total parameters and 36 billion active parameters, supporting a 32K context window
Downloads 100
Release Time : 3/26/2024

Model Overview

DBRX is a decoder-only large language model based on the Transformer architecture, employing a fine-grained Mixture of Experts architecture, suitable for text completion and code generation tasks

Model Features

Fine-grained Mixture of Experts Architecture
Adopts a 16-select-4 expert structure, increasing expert combination possibilities by 65 times compared to traditional 8-select-2 architectures
Large-scale Pretraining
Trained on 12 trillion carefully selected tokens, with data quality improved by 2 times compared to previous generations
Efficient Inference
Optimizes computational efficiency through GQA grouped query attention mechanism and FlashAttention2
Enterprise-grade Deployment Support
Can be deployed in production environments via Databricks Foundation Model API

Model Capabilities

English text completion
Code generation and completion
Long-context processing (32K tokens)
Commercial scenario fine-tuning

Use Cases

General text processing
Document auto-continuation
Generates coherent subsequent content based on the opening paragraph
Software development
Code auto-completion
Generates implementation code based on function signatures or comments
Excellent performance on HumanEval benchmark
Enterprise applications
Domain knowledge Q&A system
Builds professional domain Q&A assistants based on RAG architecture
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase