Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Think Twice Before Trusting: Self-Detection for Large Language Models through Comprehensive Answer Reflection (2403.09972v3)

Published 15 Mar 2024 in cs.CL

Abstract: Self-detection for LLMs seeks to evaluate the trustworthiness of the LLM's output by leveraging its own capabilities, thereby alleviating the issue of output hallucination. However, existing self-detection approaches only retrospectively evaluate answers generated by LLM, typically leading to the over-trust in incorrectly generated answers. To tackle this limitation, we propose a novel self-detection paradigm that considers the comprehensive answer space beyond LLM-generated answers. It thoroughly compares the trustworthiness of multiple candidate answers to mitigate the over-trust in LLM-generated incorrect answers. Building upon this paradigm, we introduce a two-step framework, which firstly instructs LLM to reflect and provide justifications for each candidate answer, and then aggregates the justifications for comprehensive target answer evaluation. This framework can be seamlessly integrated with existing approaches for superior self-detection. Extensive experiments on six datasets spanning three tasks demonstrate the effectiveness of the proposed framework.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Moxin Li (13 papers)
  2. Wenjie Wang (150 papers)
  3. Fuli Feng (143 papers)
  4. Fengbin Zhu (19 papers)
  5. Qifan Wang (129 papers)
  6. Tat-Seng Chua (359 papers)
Citations (8)