A fully open-source family of reasoning models built using a dataset derived by distilling DeepSeek-R1.
7b
32b
512.7K Pulls Updated 5 weeks ago
Updated 6 weeks ago
6 weeks ago
aaffe05a5e2e · 4.7GB
model
archqwen2
·
parameters7.62B
·
quantizationQ4_K_M
4.7GB
template
{{- range $i, $_ := .Messages }}
{{- $last := eq (len (slice $.Messages $i)) 1 -}}
<|im_start|>{{ .R
275B
system
Your role as an assistant involves thoroughly exploring questions through a systematic long thinking
1.3kB
license
Apache License
Version 2.0, January 2004
http://www.apache.org/lice
10kB
Readme
OpenThinker is a family of fine-tuned models from Qwen2.5 on the OpenThoughts-114k dataset, surpassing DeepSeek-R1 distillation models on some benchmarks.
The dataset is derived by distilling DeepSeek-R1 using the data pipeline available on github. More info about the dataset can be found on the dataset card at OpenThoughts-114k dataset.
Models
32B
ollama run openthinker:32b
Model Name | Dataset Size | AIME24 I/II | AIME25 I | MATH500 | GPQA Diamond | LCBv2 |
---|---|---|---|---|---|---|
LIMO-32B | 0.8k | 56.7 | 49.3 | 86.6 | 58.1 | 60.0 |
s1-32B | 1k | 36.0 | 25.3 | 84.8 | 50.5 | 40.9 |
s1.1-32B | 1k | 64.7 | 49.3 | 89.0 | 60.1 | 65.5 |
DeepSeek-R1-Distill-Qwen-32B | 800k (closed) | 76.7 | 55.9 | 89.4 | 57.6 | 71.2 |
OpenThinker-32B | 114k | 66.0 | 53.3 | 90.6 | 61.6 | 68.9 |
7B (default)
ollama run openthinker:7b
AIME24 | MATH500 | GPQA-Diamond | LCBv2 Easy | LCBv2 Medium | LCBv2 Hard | LCBv2 All | |
---|---|---|---|---|---|---|---|
OpenThinker-7B | 31.3 | 83.0 | 42.4 | 75.3 | 28.6 | 6.5 | 39.9 |
Bespoke-Stratos-7B | 22.7 | 79.6 | 38.9 | 71.4 | 25.2 | 0.8 | 35.8 |
DeepSeek-R1-Distill-Qwen-7B | 60 | 88.2 | 46.9 | 79.7 | 45.1 | 14.6 | 50.1 |
gpt-4o-0513 | 8.7 | 75.8 | 46.5 | 87.4 | 42.7 | 8.9 | 50.5 |
o1-mini | 64 | 85.6 | 60 | 92.8 | 74.7 | 39.8 | 72.8 |