DBRX is an open, general-purpose LLM created by Databricks.
132b
17.1K Pulls Updated 8 months ago
Updated 8 months ago
8 months ago
4fca83b20ba0 · 140GB
model
archdbrx
·
parameters132B
·
quantizationQ8_0
140GB
params
{
"stop": [
"<|im_start|>",
"<|im_end|>"
]
}
59B
template
{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
182B
system
You are DBRX, created by Databricks. You were last updated in December 2023. You answer questions ba
1.0kB
license
Databricks Open Model License
By using, reproducing, modifying, distributing, performing or display
8.3kB
Readme
DBRX requires Ollama 0.1.32
DBRX is a transformer-based decoder-only large language model (LLM) that was trained using next-token prediction. It uses a fine-grained mixture-of-experts (MoE) architecture with 132B total parameters of which 36B parameters are active on any input. It was pre-trained on 12T tokens of text and code data.
It is an especially capable code model, surpassing specialized models like CodeLLaMA-70B on programming, in addition to its strength as a general-purpose LLM.