Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enhancing Confidence Expression in Large Language Models Through Learning from Past Experience (2404.10315v1)

Published 16 Apr 2024 in cs.CL

Abstract: LLMs have exhibited remarkable performance across various downstream tasks, but they may generate inaccurate or false information with a confident tone. One of the possible solutions is to empower the LLM confidence expression capability, in which the confidence expressed can be well-aligned with the true probability of the generated answer being correct. However, leveraging the intrinsic ability of LLMs or the signals from the output logits of answers proves challenging in accurately capturing the response uncertainty in LLMs. Therefore, drawing inspiration from cognitive diagnostics, we propose a method of Learning from Past experience (LePe) to enhance the capability for confidence expression. Specifically, we first identify three key problems: (1) How to capture the inherent confidence of the LLM? (2) How to teach the LLM to express confidence? (3) How to evaluate the confidence expression of the LLM? Then we devise three stages in LePe to deal with these problems. Besides, to accurately capture the confidence of an LLM when constructing the training data, we design a complete pipeline including question preparation and answer sampling. We also conduct experiments using the Llama family of LLMs to verify the effectiveness of our proposed method on four datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. Gpt-4 technical report. arXiv preprint arXiv:2303.08774.
  2. Gwern Branwen. 2020. Gpt-3 nonfiction- calibration. Technical report, The institution that published. Last accessed on 2022-04-24.
  3. Training verifiers to solve math word problems. arXiv preprint arXiv:2110.14168.
  4. Yarin Gal and Zoubin Ghahramani. 2016. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In international conference on machine learning, pages 1050–1059. PMLR.
  5. Tyler Giallanza and Declan Iain Campbell. 2024. Context-sensitive semantic reasoning in large language models. In ICLR 2024 Workshop on Representational Alignment.
  6. Xiezhi: An ever-updating benchmark for holistic domain knowledge evaluation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pages 18099–18107.
  7. How close is chatgpt to human experts? comparison corpus, evaluation, and detection. arXiv preprint arXiv:2301.07597.
  8. Small language model can self-correct. In AAAI Conference on Artificial Intelligence.
  9. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. ArXiv, abs/2311.05232.
  10. C-eval: A multi-level multi-discipline chinese evaluation suite for foundation models. Advances in Neural Information Processing Systems, 36.
  11. Language models (mostly) know what they know. ArXiv, abs/2207.05221.
  12. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361.
  13. Inference-time intervention: Eliciting truthful answers from a language model. Advances in Neural Information Processing Systems, 36.
  14. Think twice before assure: Confidence estimation for large language models through reflection on multiple answers. arXiv preprint arXiv:2403.09972.
  15. Teaching models to express their uncertainty in words. arXiv preprint arXiv:2205.14334.
  16. M3ke: A massive multi-level multi-subject knowledge evaluation benchmark for chinese large language models. arXiv preprint arXiv:2305.10263.
  17. Survey of computerized adaptive testing: A machine learning perspective. arXiv preprint arXiv:2404.00712.
  18. Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101.
  19. Can a suit of armor conduct electricity? a new dataset for open book question answering. arXiv preprint arXiv:1809.02789.
  20. Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35:27730–27744.
  21. Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36.
  22. Investigating the factual knowledge boundary of large language models with retrieval augmentation. arXiv preprint arXiv:2307.11019.
  23. Retrieval augmentation reduces hallucination in conversation. arXiv preprint arXiv:2104.07567.
  24. Beyond the imitation game: Quantifying and extrapolating the capabilities of language models. arXiv preprint arXiv:2206.04615.
  25. Learning to summarize with human feedback. Advances in Neural Information Processing Systems, 33:3008–3021.
  26. Commonsenseqa: A question answering challenge targeting commonsense knowledge. arXiv preprint arXiv:1811.00937.
  27. Fine-tuning language models for factuality. arXiv preprint arXiv:2311.08401.
  28. Just ask for calibration: Strategies for eliciting calibrated confidence scores from language models fine-tuned with human feedback. arXiv preprint arXiv:2305.14975.
  29. "my answer is c": First-token probabilities do not match text answers in instruction-tuned language models. ArXiv, abs/2402.14499.
  30. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171.
  31. Chain of thought prompting elicits reasoning in large language models. ArXiv, abs/2201.11903.
  32. Can llms express their uncertainty? an empirical evaluation of confidence elicitation in llms. arXiv preprint arXiv:2306.13063.
  33. Do large language models know what they don’t know? arXiv preprint arXiv:2305.18153.
  34. Navigating the grey area: How expressions of uncertainty and overconfidence affect language models. arXiv preprint arXiv:2302.13439.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Haixia Han (4 papers)
  2. Tingyun Li (2 papers)
  3. Shisong Chen (4 papers)
  4. Jie Shi (32 papers)
  5. Chengyu Du (15 papers)
  6. Yanghua Xiao (151 papers)
  7. Jiaqing Liang (62 papers)
  8. Xin Lin (81 papers)
Citations (5)