HALO: An Ontology for Representing and Categorizing Hallucinations in Large Language Models (2312.05209v2)
Abstract: Recent progress in generative AI, including LLMs like ChatGPT, has opened up significant opportunities in fields ranging from natural language processing to knowledge discovery and data mining. However, there is also a growing awareness that the models can be prone to problems such as making information up or `hallucinations', and faulty reasoning on seemingly simple problems. Because of the popularity of models like ChatGPT, both academic scholars and citizen scientists have documented hallucinations of several different types and severity. Despite this body of work, a formal model for describing and representing these hallucinations (with relevant meta-data) at a fine-grained level, is still lacking. In this paper, we address this gap by presenting the Hallucination Ontology or HALO, a formal, extensible ontology written in OWL that currently offers support for six different types of hallucinations known to arise in LLMs, along with support for provenance and experimental metadata. We also collect and publish a dataset containing hallucinations that we inductively gathered across multiple independent Web sources, and show that HALO can be successfully used to model this dataset and answer competency questions.
- Friend of a friend with benefits ontology (foaf+): extending a social network ontology for public health. BMC Medical Informatics and Decision Making, 20(10):1–14, 2020.
- The international data spaces information model–an ontology for sovereign exchange of digital content. In International Semantic Web Conference, pages 176–192. Springer, 2020.
- A multitask, multilingual, multimodal evaluation of chatgpt on reasoning, hallucination, and interactivity, 2023.
- D. Berrueta and J. Phipps. Best practice recipes for publishing rdf vocabularies, 08 2008.
- Sparks of artificial general intelligence: Early experiments with gpt-4, 2023.
- Mireot: the minimum information to reference an external ontology term. Applied Ontology, 6:23–33, 01 2011.
- Chatlaw: Open-source legal large language model with integrated external knowledge bases, 2023.
- An upper ontology for modern science branches and related entities. In The Semantic Web: 20th International Conference, ESWC 2023, Hersonissos, Crete, Greece, May 28–June 1, 2023, Proceedings, page 436–453, Berlin, Heidelberg, 2023. Springer-Verlag.
- Towards an ontology for urban tourism. In Proceedings of the 36th annual ACM symposium on applied computing, pages 1887–1890, 2021.
- D. Garijo. Widoco: A wizard for documenting ontologies. In C. d’Amato, M. Fernandez, V. Tamma, F. Lecue, P. Cudré-Mauroux, J. Sequeda, C. Lange, and J. Heflin, editors, The Semantic Web – ISWC 2017, pages 94–102, Cham, 2017. Springer International Publishing.
- D. Garijo and M. Osorio. Oba: An ontology-based framework for creating rest apis for knowledge graphs. In The Semantic Web–ISWC 2020: 19th International Semantic Web Conference, Athens, Greece, November 2–6, 2020, Proceedings, Part II 19, pages 48–64. Springer, 2020.
- Hallucinations in large multilingual translation models, 2023.
- A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions, 2023.
- Dehallucinating large language models using formal methods guided iterative prompting. In 2023 IEEE International Conference on Assured Autonomy (ICAA), pages 149–152, 2023.
- Survey of hallucination in natural language generation. ACM Comput. Surv., 55(12), mar 2023.
- Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension, 2019.
- An ontology for the materials design domain. In International Semantic Web Conference, pages 212–227. Springer, 2020.
- Halueval: A large-scale hallucination evaluation benchmark for large language models, 2023.
- Competition-level code generation with alphacode. Science, 378(6624):1092–1097, 2022.
- Selfcheckgpt: Zero-resource black-box hallucination detection for generative large language models, 2023.
- M. A. Musen. The protégé project: A look back and a look forward. AI Matters, 1(4):4–12, jun 2015.
- Self-contradictory hallucinations of large language models: Evaluation, detection and mitigation, 2023.
- OpenAI. Gpt-4 technical report, 2023.
- OOPS! (OntOlogy Pitfall Scanner!): An On-line Tool for Ontology Evaluation. International Journal on Semantic Web and Information Systems (IJSWIS), 10(2):7–34, 2014.
- Lot: An industrial oriented ontology engineering framework. Engineering Applications of Artificial Intelligence, 111:104755, 2022.
- Zero-shot text-to-image generation, 2021.
- A concise ontology to support research on complex, multimodal clinical reasoning. In The Semantic Web: 20th International Conference, ESWC 2023, Hersonissos, Crete, Greece, May 28–June 1, 2023, Proceedings, page 390–407, Berlin, Heidelberg, 2023. Springer-Verlag.
- A survey of hallucination in large foundation models, 2023.
- On accurate evaluation of gans for language generation, 2019.
- Towards expert-level medical question answering with large language models, 2023.
- Lamda: Language models for dialog applications, 2022.
- A stitch in time saves nine: Detecting and mitigating hallucinations of llms by validating low-confidence generation, 2023.
- Cognitive mirage: A review of hallucinations in large language models, 2023.
- Biobart: Pretraining and evaluation of a biomedical generative language model, 2022.
- Lion: Latent point diffusion models for 3d shape generation, 2022.
- How language model hallucinations can snowball, 2023.
- Siren’s song in the ai ocean: A survey on hallucination in large language models, 2023.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.