This is a 32B reasoning model trained from Qwen2.5-32B-Instruct with 17K data. The performance is on par with o1-preview model on both math and coding.

240 2 months ago

2 months ago

ba1e6161ae00 · 20GB

model
qwen2
·
32.8B
·
Q4_K_M

Readme

https://huggingface.co/NovaSky-AI/Sky-T1-32B-Preview

Model Description

This is a 32B reasoning model trained from Qwen2.5-32B-Instruct with 17K data. The performance is on par with o1-preview model on both math and coding. Please see our blog post for more details.

Developed by: NovaSky Team from Sky Computing Lab at UC Berkeley.

Training Details

Training Data

17K verified correct responses from Qwen/QwQ-32B-Preview on coding, math. In addition, we add the science portion from the Still-2 paper.

@misc{sky_t1_2025, author = {NovaSky Team}, title = {Sky-T1: Fully open-source reasoning model with o1-preview performance in $450 budget}, howpublished = {https://novasky-ai.github.io/posts/sky-t1}, note = {Accessed: 2025-01-09}, year = {2025} }