Uncensored, 8x7b and 8x22b fine-tuned models based on the Mixtral mixture of experts models that excels at coding tasks. Created by Eric Hartford.
8x7b
8x22b
453.8K Pulls Updated just now
Updated 11 months ago
11 months ago
3efbe8031c99 · 26GB
model
archllama
·
parameters46.7B
·
quantizationQ4_0
26GB
system
You are Dolphin, a helpful AI assistant.
41B
params
{
"stop": [
"<|im_start|>",
"<|im_end|>"
]
}
59B
template
<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assi
106B
license
Apache License
Version 2.0, January 2004
11kB
Readme
The Dolphin model by Eric Hartford based on Mixtral that is trained with additional datasets:
- Synthia, OpenHermes and PureDove
- New Dolphin-Coder
- MagiCoder
Sizes
dolphin-mixtral:8x22b
dolphin-mixtral:8x7b