Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Let's Sample Step by Step: Adaptive-Consistency for Efficient Reasoning and Coding with LLMs (2305.11860v2)

Published 19 May 2023 in cs.CL

Abstract: A popular approach for improving the correctness of output from LLMs is Self-Consistency - poll the LLM multiple times and output the most frequent solution. Existing Self-Consistency techniques always generate a constant number of samples per question, where a better approach will be to non-uniformly distribute the available budget based on the amount of agreement in the samples generated so far. In response, we introduce Adaptive-Consistency, a cost-efficient, model-agnostic technique that dynamically adjusts the number of samples per question using a lightweight stopping criterion. Our experiments over 17 reasoning and code generation datasets and three LLMs demonstrate that Adaptive-Consistency reduces sample budget by up to 7.9 times with an average accuracy drop of less than 0.1%. Our code and data are available at https://www.sample-step-by-step.info

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Pranjal Aggarwal (9 papers)
  2. Aman Madaan (30 papers)
  3. Yiming Yang (151 papers)
  4. Mausam (69 papers)
Citations (16)

Summary

We haven't generated a summary for this paper yet.