Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enhancing Large Language Models with Pseudo- and Multisource- Knowledge Graphs for Open-ended Question Answering (2402.09911v1)

Published 15 Feb 2024 in cs.CL and cs.AI

Abstract: Mitigating the hallucinations of LLMs and enhancing them is a crucial task. Although some existing methods employ model self-enhancement techniques, they fall short of effectively addressing unknown factual hallucinations. Using Knowledge Graph (KG) enhancement approaches fails to address the generalization across different KG sources and the enhancement of open-ended answer questions simultaneously. To tackle these limitations, there is a framework that combines Pseudo-Graph Generation and Atomic Knowledge Verification proposed. The enhancement of LLM using KG in an open-ended question-answering setting is implemented by leveraging the Pseudo-Graph Generation. Atomic Knowledge Verification utilizes atomic-level knowledge querying and verification to achieve generalizability under different KG sources. Compared to the baseline, this approach yields a minimum improvement of 11.5 in the ROUGE-L score for open-ended questions. For precise questions, we observe a minimum accuracy improvement of 7.5. Moreover, there is also demonstration that this framework exhibits generalizability across different KG sources. In summary, our results pave the way for enhancing LLMs by incorporating Pseudo- and Multisource-KGs, particularly in the context of open-ended questions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Anonymous. 2024. Think-on-graph: Deep and responsible reasoning of large language model on knowledge graph. In The Twelfth International Conference on Learning Representations.
  2. Large-scale simple question answering with memory networks. ArXiv, abs/1506.02075.
  3. Large-scale simple question answering with memory networks.
  4. Translating embeddings for modeling multi-relational data. In Advances in Neural Information Processing Systems, volume 26. Curran Associates, Inc.
  5. Sparql-qa enters the qald challenge. In NLIWoD@ESWC.
  6. Language models are few-shot learners. ArXiv, abs/2005.14165.
  7. Shuaichen Chang and Eric Fosler-Lussier. 2023. How to prompt llms for text-to-sql: A study in zero-shot, single-domain, and cross-domain settings. arXiv preprint arXiv:2305.11853, abs/2305.11853.
  8. Palm: Scaling language modeling with pathways. ArXiv, abs/2204.02311.
  9. Mitigating large language model hallucinations via autonomous knowledge graph-based retrofitting.
  10. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions.
  11. Large language models are zero-shot reasoners. In Advances in Neural Information Processing Systems.
  12. Retrieval-augmented generation for knowledge-intensive nlp tasks. In Advances in Neural Information Processing Systems, volume 33, pages 9459–9474. Curran Associates, Inc.
  13. To copy rather than memorize: A vertical learning paradigm for knowledge graph completion. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 6335–6347, Toronto, Canada. Association for Computational Linguistics.
  14. Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74–81, Barcelona, Spain. Association for Computational Linguistics.
  15. Kam-cot: Knowledge augmented multimodal chain-of-thoughts reasoning. ArXiv, abs/2401.12863.
  16. OpenAI. 2023. Gpt-4 technical report. ArXiv, abs/2303.08774.
  17. Training language models to follow instructions with human feedback. ArXiv, abs/2203.02155.
  18. Unifying large language models and knowledge graphs: A roadmap. IEEE Transactions on Knowledge and Data Engineering.
  19. Graph-guided reasoning for multi-hop question answering in large language models.
  20. Qald-9-plus: A multilingual dataset for question answering over dbpedia and wikidata translated by native speakers. In 2022 IEEE 16th International Conference on Semantic Computing (ICSC), pages 229–234.
  21. Similarity-weighted construction of contextualized commonsense knowledge graphs for knowledge-intense argumentation tasks. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 6130–6158, Toronto, Canada. Association for Computational Linguistics.
  22. Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence embeddings using Siamese BERT-networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3982–3992, Hong Kong, China. Association for Computational Linguistics.
  23. Sql-palm: Improved large language modeladaptation for text-to-sql. arXiv preprint arXiv:2306.00739, abs/2306.00739.
  24. Llama 2: Open foundation and fine-tuned chat models. ArXiv, abs/2307.09288.
  25. Knowledge-driven cot: Exploring faithful reasoning in llms for knowledge-intensive question answering. ArXiv, abs/2308.13259.
  26. Self-consistency improves chain of thought reasoning in language models. In The Eleventh International Conference on Learning Representations.
  27. Chain-of-thought prompting elicits reasoning in large language models. In Advances in Neural Information Processing Systems, volume 35, pages 24824–24837. Curran Associates, Inc.
  28. Cognitive mirage: A review of hallucinations in large language models.
  29. Evaluating the code quality of ai-assisted code generation tools: An empirical study on github copilot, amazon codewhisperer, and chatgpt.
  30. Automatic chain of thought prompting in large language models. In The Eleventh International Conference on Learning Representations.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jiaxiang Liu (39 papers)
  2. Tong Zhou (124 papers)
  3. Yubo Chen (58 papers)
  4. Kang Liu (207 papers)
  5. Jun Zhao (469 papers)
Citations (2)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets