Skip to content

jonathanyin12/distilled-reasoning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Reasoning with Less: Distilling Reasoning Models with Small Datasets

Paper

model_score_vs_data_size

Figure 1: Pass@1 accuracy on AIME 2024–2025 vs. the number of SFT examples for various 32B parameter distilled reasoning models. Notably, QwQ-32B-Distill-32B from our work reaches 63.5% average accuracy using <1K examples, rivaling or exceeding many state-of-the-art large-data baselines. See Table 2 for more details.

teacher_student_comparison_aime2025 Figure 2: Pass@1 accuracy of student models versus their teacher model’s accuracy on AIME 2025. Each marker represents a specific teacher–student pairing (marker shape denotes the teacher model; color denotes student scale: 1.5B in blue, 7B in orange, 32B in green). Larger students exhibit a much stronger positive correlation between teacher and student performance.

Setup

python3.12 -m venv env
source env/bin/activate
pip install -r requirements.txt

About

Efficiently distilling reasoning ability into non‑reasoning LLMs

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published