Papers
Topics
Authors
Recent
2000 character limit reached

MetaCheckGPT -- A Multi-task Hallucination Detector Using LLM Uncertainty and Meta-models (2404.06948v2)

Published 10 Apr 2024 in cs.CL and cs.AI

Abstract: Hallucinations in LLMs have recently become a significant problem. A recent effort in this direction is a shared task at Semeval 2024 Task 6, SHROOM, a Shared-task on Hallucinations and Related Observable Overgeneration Mistakes. This paper describes our winning solution ranked 1st and 2nd in the 2 sub-tasks of model agnostic and model aware tracks respectively. We propose a meta-regressor framework of LLMs for model evaluation and integration that achieves the highest scores on the leaderboard. We also experiment with various transformer-based models and black box methods like ChatGPT, Vectara, and others. In addition, we perform an error analysis comparing GPT4 against our best model which shows the limitations of the former.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (24)
  1. Gpt-4 technical report. Technical report, OpenAI.
  2. Amos Azaria and Tom Mitchell. 2023. The internal state of an llm knows when its lying. arXiv preprint arXiv:2304.13734.
  3. Generationary or “how we went beyond word sense inventories and learned to gloss”. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 7207–7221, Online. Association for Computational Linguistics.
  4. Detecting and mitigating hallucinations in machine translation: Model internal workings alone do well, sentence similarity even better.
  5. Hallucinations in large multilingual translation models.
  6. Looking for a needle in a haystack: A comprehensive study of hallucinations in neural machine translation.
  7. Deberta: Decoding-enhanced bert with disentangled attention.
  8. The factual inconsistency problem in abstractive text summarization: A survey.
  9. Simon Hughes. 2023. Cut the bull… detecting hallucinations in large language models.
  10. Mixtral of experts.
  11. Chain of natural language inference for reducing large language model ungrounded hallucinations. arXiv, cs.CL(arXiv:2310.03951).
  12. Truthfulqa: Measuring how models mimic human falsehoods.
  13. Roberta: A robustly optimized bert pretraining approach.
  14. Selfcheckgpt: Zero-resource black-box hallucination detection for generative large language models.
  15. On faithfulness and factuality in abstractive summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1906–1919, Online. Association for Computational Linguistics.
  16. Semeval-2024 shared task 6: Shroom, a shared-task on hallucinations and related observable overgeneration mistakes. In Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024), pages 1980–1994, Mexico City, Mexico. Association for Computational Linguistics.
  17. Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter.
  18. Retrieval augmentation reduces hallucination in conversation. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 3784–3803, Punta Cana, Dominican Republic. Association for Computational Linguistics.
  19. Retrieval augmentation reduces hallucination in conversation. arXiv, cs.CL(arXiv:2104.07567).
  20. Just ask for calibration: Strategies for eliciting calibrated confidence scores from language models fine-tuned with human feedback. https://doi.org/10.48550/arXiv.2305.14975. ArXiv:2305.14975v2 [cs.CL].
  21. Llama 2: Open foundation and fine-tuned chat models.
  22. Mutual information alleviates hallucinations in abstractive summarization.
  23. A stitch in time saves nine: Detecting and mitigating hallucinations of llms by validating low-confidence generation.
  24. Yijun Xiao and William Yang Wang. 2021. On hallucination and predictive uncertainty in conditional language generation.

Summary

We haven't generated a summary for this paper yet.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

Sign up for free to view the 1 tweet with 1 like about this paper.