Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Explanations from Large Language Models Make Small Reasoners Better (2210.06726v1)

Published 13 Oct 2022 in cs.CL

Abstract: Integrating free-text explanations to in-context learning of LLMs (LLM) is shown to elicit strong reasoning capabilities along with reasonable explanations. In this paper, we consider the problem of leveraging the explanations generated by LLM to improve the training of small reasoners, which are more favorable in real-production deployment due to their low cost. We systematically explore three explanation generation approaches from LLM and utilize a multi-task learning framework to facilitate small models to acquire strong reasoning power together with explanation generation capabilities. Experiments on multiple reasoning tasks show that our method can consistently and significantly outperform finetuning baselines across different settings, and even perform better than finetuning/prompting a 60x larger GPT-3 (175B) model by up to 9.5% in accuracy. As a side benefit, human evaluation further shows that our method can generate high-quality explanations to justify its predictions, moving towards the goal of explainable AI.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Shiyang Li (24 papers)
  2. Jianshu Chen (66 papers)
  3. Yelong Shen (83 papers)
  4. Zhiyu Chen (60 papers)
  5. Xinlu Zhang (15 papers)
  6. Zekun Li (73 papers)
  7. Hong Wang (254 papers)
  8. Jing Qian (81 papers)
  9. Baolin Peng (72 papers)
  10. Yi Mao (78 papers)
  11. Wenhu Chen (134 papers)
  12. Xifeng Yan (52 papers)
Citations (112)