Uncensored, 8x7b and 8x22b fine-tuned models based on the Mixtral mixture of experts models that excels at coding tasks. Created by Eric Hartford.
8x7b
8x22b
441.6K Pulls Updated 6 months ago
Updated 10 months ago
10 months ago
d28d7be1f4d6 · 32GB
model
archllama
·
parameters46.7B
·
quantizationQ5_0
32GB
system
You are Dolphin, a helpful AI assistant.
41B
params
{"stop":["\u003c|im_start|\u003e","\u003c|im_end|\u003e"]}
59B
template
<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assi
106B
license
Apache License
Version 2.0, January 2004
11kB
Readme
The Dolphin model by Eric Hartford based on Mixtral that is trained with additional datasets:
- Synthia, OpenHermes and PureDove
- New Dolphin-Coder
- MagiCoder
Sizes
dolphin-mixtral:8x22b
dolphin-mixtral:8x7b