Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

KnowGPT: Knowledge Graph based Prompting for Large Language Models (2312.06185v5)

Published 11 Dec 2023 in cs.CL and cs.AI

Abstract: LLMs have demonstrated remarkable capabilities in many real-world applications. Nonetheless, LLMs are often criticized for their tendency to produce hallucinations, wherein the models fabricate incorrect statements on tasks beyond their knowledge and perception. To alleviate this issue, researchers have explored leveraging the factual knowledge in knowledge graphs (KGs) to ground the LLM's responses in established facts and principles. However, most state-of-the-art LLMs are closed-source, making it challenging to develop a prompting framework that can efficiently and effectively integrate KGs into LLMs with hard prompts only. Generally, existing KG-enhanced LLMs usually suffer from three critical issues, including huge search space, high API costs, and laborious prompt engineering, that impede their widespread application in practice. To this end, we introduce a novel Knowledge Graph based PrompTing framework, namely KnowGPT, to enhance LLMs with domain knowledge. KnowGPT contains a knowledge extraction module to extract the most informative knowledge from KGs, and a context-aware prompt construction module to automatically convert extracted knowledge into effective prompts. Experiments on three benchmarks demonstrate that KnowGPT significantly outperforms all competitors. Notably, KnowGPT achieves a 92.6% accuracy on OpenbookQA leaderboard, comparable to human-level performance.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Ai unreliable answers: A case study on chatgpt. In ICHCI, pp.  23–40. Springer, 2023.
  2. Olivier Bodenreider. The Unified Medical Language System (UMLS): integrating biomedical terminology. Nucleic Acids Research, 32:D267–D270, 01 2004. ISSN 0305-1048.
  3. Freebase: a collaboratively created graph database for structuring human knowledge. In Proceedings of the 2008 ACM SIGMOD international conference on Management of data, pp.  1247–1250. ACM, 2008.
  4. Language models are few-shot learners. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp.  1877–1901, 2020.
  5. A new algorithm for non-stationary contextual bandits: Efficient, optimal and parameter-free. In COLT. PMLR, 2019.
  6. From ‘f’to ‘a’on the ny regents science exams: An overview of the aristo project. AI Magazine, 41(4):39–53, 2020.
  7. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  8. Hierarchy-aware multi-hop question answering over knowledge graphs. In WWW, pp.  2519–2527, 2023a.
  9. Active ensemble learning for knowledge graph error detection. In WSDM, pp.  877–885, 2023b.
  10. Scalable multi-hop relational reasoning for knowledge-aware question answering. In EMNLP, pp.  1295–1309, 2020.
  11. Learning to fake it: limited responses and fabricated references provided by chatgpt for medical questions. Mayo Clinic Proceedings: Digital Health, 1(3):226–234, 2023.
  12. Mvp-tuning: Multi-view knowledge retrieval with prompt tuning for commonsense reasoning. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp.  13417–13432, 2023.
  13. Clues before answers: Generation-enhanced multiple-choice qa. arXiv preprint arXiv:2205.00274, 2022.
  14. X-factr: Multilingual factual knowledge retrieval from pretrained language models. arXiv preprint arXiv:2010.06189, 2020.
  15. What disease does this patient have? a large-scale open domain question answering dataset from medical exams. Applied Sciences, 11(14), 2021. ISSN 2076-3417. doi: 10.3390/app11146421. URL https://www.mdpi.com/2076-3417/11/14/6421.
  16. Unifiedqa: Crossing format boundaries with a single qa system. In EMNLP 2020, pp.  1896–1907, 2020.
  17. Performance of chatgpt on usmle: Potential for ai-assisted medical education using large language models. PLoS digital health, 2(2):e0000198, 2023.
  18. The power of scale for parameter-efficient prompt tuning. arXiv preprint arXiv:2104.08691, 2021.
  19. Prefix-tuning: Optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190, 2021.
  20. Kagnet: Knowledge-aware graph networks for commonsense reasoning. In EMNLP-IJCNLP, pp.  2829–2839, 2019.
  21. Self-alignment pretraining for biomedical entity representations. arXiv preprint arXiv:2010.11784, 2020a.
  22. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9):1–35, 2023.
  23. K-bert: Enabling language representation with knowledge graph. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pp.  2901–2908, 2020b.
  24. P-tuning: Prompt tuning can be comparable to fine-tuning across scales and tasks. In ACL, pp.  61–68, 2022.
  25. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
  26. Can a suit of armor conduct electricity? a new dataset for open book question answering. In EMNLP, pp.  2381–2391, 2018.
  27. OpenAI. Gpt-4 technical report, 2023.
  28. Unifying large language models and knowledge graphs: A roadmap. arXiv preprint arXiv:2306.08302, 2023.
  29. Language models as knowledge bases? arXiv preprint arXiv:1909.01066, 2019.
  30. Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100, 2022.
  31. In chatgpt we trust? measuring and characterizing the reliability of chatgpt. arXiv preprint arXiv:2304.08979, 2023.
  32. Conceptnet 5.5: An open multilingual graph of general knowledge. In Proceedings of the AAAI conference on artificial intelligence, volume 31, 2017.
  33. Cokebert: Contextual knowledge selection and embedding towards enhanced pre-trained language models. AI Open, 2:127–134, 2021.
  34. YAGO: a core of semantic knowledge. In Proceedings of the 16th international conference on World Wide Web, pp.  697–706. ACM, 2007.
  35. Colake: Contextualized language and knowledge embedding. arXiv preprint arXiv:2010.00309, 2020.
  36. Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation. arXiv preprint arXiv:2107.02137, 2021a.
  37. Jointlk: Joint reasoning with language models and knowledge graphs for commonsense question answering. arXiv preprint arXiv:2112.02732, 2021b.
  38. Reinforcement learning: An introduction. MIT press, 2018.
  39. Commonsenseqa: A question answering challenge targeting commonsense knowledge. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp.  4149–4158, 2019.
  40. Grapeqa: Graph augmentation and pruning to enhance question-answering. In Companion Proceedings of the ACM Web Conference 2023, pp.  1138–1144, 2023.
  41. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  42. Exploring compact reinforcement-learning representations with linear regression. arXiv, 2012.
  43. Gnn is a counter? revisiting gnn for question answering, 2021.
  44. DrugBank 5.0: a major update to the DrugBank database for 2018. Nucleic Acids Research, 46:D1074–D1082, 2017.
  45. Deeppath: A reinforcement learning method for knowledge graph reasoning. In EMNLP, pp.  564–573, 2017.
  46. Qa-gnn: Reasoning with language models and knowledge graphs for question answering. NAACL, 2021.
  47. Deep bidirectional language-knowledge graph pretraining, 2022.
  48. Data-centric artificial intelligence: A survey. arXiv preprint arXiv:2303.10158, 2023.
  49. Greaselm: Graph reasoning enhanced language models for question answering. arXiv preprint arXiv:2201.08860, 2022.
  50. Ernie: Enhanced language representation with informative entities. arXiv preprint arXiv:1905.07129, 2019.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Qinggang Zhang (19 papers)
  2. Junnan Dong (14 papers)
  3. Hao Chen (1005 papers)
  4. Xiao Huang (112 papers)
  5. Daochen Zha (56 papers)
  6. Zailiang Yu (1 paper)
Citations (1)