Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Confidence Calibration and Rationalization for LLMs via Multi-Agent Deliberation (2404.09127v3)

Published 14 Apr 2024 in cs.CL

Abstract: Uncertainty estimation is a significant issue for current LLMs that are generally poorly calibrated and over-confident, especially with reinforcement learning from human feedback (RLHF). Unlike humans, whose decisions and confidences not only stem from intrinsic beliefs but can also be adjusted through daily observations, existing calibration methods for LLMs focus on estimating or eliciting individual confidence without taking full advantage of the "Collective Wisdom": the interaction among multiple LLMs that can collectively improve both accuracy and calibration. In this work, we propose Collaborative Calibration, a post-hoc training-free calibration strategy that leverages the collaborative and expressive capabilities of multiple tool-augmented LLM agents in a simulated group deliberation process. We demonstrate the effectiveness of Collaborative Calibration on generative QA tasks across various domains, showing its potential in harnessing the rationalization of collectively calibrated confidence assessments and improving the reliability of model predictions.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Ruixin Yang (15 papers)
  2. Shirley Anugrah Hayati (13 papers)
  3. Bin Hu (217 papers)
  4. Dongyeop Kang (72 papers)
  5. Dheeraj Rajagopal (20 papers)
Citations (2)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets