Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ConU: Conformal Uncertainty in Large Language Models with Correctness Coverage Guarantees (2407.00499v3)

Published 29 Jun 2024 in cs.CL, cs.AI, and cs.LG

Abstract: Uncertainty quantification (UQ) in natural language generation (NLG) tasks remains an open challenge, exacerbated by the closed-source nature of the latest LLMs. This study investigates applying conformal prediction (CP), which can transform any heuristic uncertainty notion into rigorous prediction sets, to black-box LLMs in open-ended NLG tasks. We introduce a novel uncertainty measure based on self-consistency theory, and then develop a conformal uncertainty criterion by integrating the uncertainty condition aligned with correctness into the CP algorithm. Empirical evaluations indicate that our uncertainty measure outperforms prior state-of-the-art methods. Furthermore, we achieve strict control over the correctness coverage rate utilizing 7 popular LLMs on 4 free-form NLG datasets, spanning general-purpose and medical scenarios. Additionally, the calibrated prediction sets with small size further highlights the efficiency of our method in providing trustworthy guarantees for practical open-ended NLG applications.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Zhiyuan Wang (102 papers)
  2. Jinhao Duan (23 papers)
  3. Lu Cheng (73 papers)
  4. Yue Zhang (620 papers)
  5. Qingni Wang (5 papers)
  6. Xiaoshuang Shi (20 papers)
  7. Kaidi Xu (85 papers)
  8. Hengtao Shen (16 papers)
  9. Xiaofeng Zhu (56 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.