When you purchase through links on our site, we may earn an affiliate commission.Heres how it works.
The researchers managed to achieve this milestone by distilling information from proprietary larger AI models.
Distillation is the process where a small AI model extracts information from larger AI models.

s1 AI reasoning model.
s1’s training process took less than 30 minutes using 16 NVIDIA H100 GPUs.
The model is based on Qwen2.5, an open-source Alibaba AI model.
This can lead the model to doublecheck its answer, often fixing incorrect reasoning steps, the researchers noted.

As a result, the AI model seemingly generated well-curated and accurate answers.
You cancheck out the s1 model on GitHub.








