Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LaRS: Latent Reasoning Skills for Chain-of-Thought Reasoning (2312.04684v3)

Published 7 Dec 2023 in cs.CL and cs.AI

Abstract: Chain-of-thought (CoT) prompting is a popular in-context learning (ICL) approach for LLMs, especially when tackling complex reasoning tasks. Traditional ICL approaches construct prompts using examples that contain questions similar to the input question. However, CoT prompting, which includes crucial intermediate reasoning steps (rationales) within its examples, necessitates selecting examples based on these rationales rather than the questions themselves. Existing methods require human experts or pre-trained LLMs to describe the skill, a high-level abstraction of rationales, to guide the selection. These methods, however, are often costly and difficult to scale. Instead, this paper introduces a new approach named Latent Reasoning Skills (LaRS) that employs unsupervised learning to create a latent space representation of rationales, with a latent variable called a reasoning skill. Concurrently, LaRS learns a reasoning policy to determine the required reasoning skill for a given question. Then the ICL examples are selected by aligning the reasoning skills between past examples and the question. This approach is theoretically grounded and compute-efficient, eliminating the need for auxiliary LLM inference or manual prompt design. Empirical results demonstrate that LaRS consistently outperforms SOTA skill-based selection methods, processing example banks four times faster, reducing LLM inferences during the selection stage by half, and showing greater robustness to sub-optimal example banks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zifan Xu (19 papers)
  2. Haozhu Wang (10 papers)
  3. Dmitriy Bespalov (5 papers)
  4. Peter Stone (184 papers)
  5. Yanjun Qi (68 papers)
  6. Xian Wu (139 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com