DBRX Base
LLMLLM
Base pretrained DBRX model without instruction tuning. 132B total parameters, 36B active (MoE 16 experts, top-4). Pretrained on 12T tokens, 32K context window.
Available on
Base pretrained DBRX model without instruction tuning. 132B total parameters, 36B active (MoE 16 experts, top-4). Pretrained on 12T tokens, 32K context window.
Coding★
Coding
Reasoning★
Reasoning
Long context★
Reasoning
Multilingual★
Language