Models
Turbo
Sign in
Download
Models
Download
Sign in
deepseek-v3
:671b-fp16
2.1M
Downloads
Updated
7 months ago
A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.
A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.
Cancel
671b
deepseek-v3:671b-fp16
...
/
params
f4d24e9138dd · 148B
{
"stop": [
"<|begin▁of▁sentence|>",
"<|end▁of▁sentence|>",
"<|User|>",
"<|Assistant|>"
]
}