Enhancing Reasoning Capabilities in SLMs with Reward Guided Dataset Distillation (2507.00054v1)
Abstract: The push to compress and impart the proficiency of LLMs into more deployable and efficient Small LLMs (SLMs) has benefited from improvements in knowledge distillation (KD) techniques. These techniques allow a smaller student model to learn from a more capable and larger teacher model's responses. However, distillation often revolves around the student model merely copying the teacher's in-distribution responses, limiting its generalisability. This limitation is amplified on reasoning tasks and can be computationally expensive. In this study, we propose AdvDistill, a reward-guided dataset distillation framework. We utilise multiple generations (responses) from a teacher for each prompt and assign rewards based on rule-based verifiers. These varying and normally distributed rewards serve as weights when training student models. Our methods and their subsequent behavioural analysis demonstrate a significant improvement in student model performance for mathematical and complex reasoning tasks, showcasing the efficacy and benefits of incorporating a rewarding mechanism in dataset distillation processes.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run custom paper prompts using GPT-5 on this paper.