Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

HyKGE: A Hypothesis Knowledge Graph Enhanced Framework for Accurate and Reliable Medical LLMs Responses (2312.15883v2)

Published 26 Dec 2023 in cs.CL and cs.AI

Abstract: In this paper, we investigate the retrieval-augmented generation (RAG) based on Knowledge Graphs (KGs) to improve the accuracy and reliability of LLMs. Recent approaches suffer from insufficient and repetitive knowledge retrieval, tedious and time-consuming query parsing, and monotonous knowledge utilization. To this end, we develop a Hypothesis Knowledge Graph Enhanced (HyKGE) framework, which leverages LLMs' powerful reasoning capacity to compensate for the incompleteness of user queries, optimizes the interaction process with LLMs, and provides diverse retrieved knowledge. Specifically, HyKGE explores the zero-shot capability and the rich knowledge of LLMs with Hypothesis Outputs to extend feasible exploration directions in the KGs, as well as the carefully curated prompt to enhance the density and efficiency of LLMs' responses. Furthermore, we introduce the HO Fragment Granularity-aware Rerank Module to filter out noise while ensuring the balance between diversity and relevance in retrieved knowledge. Experiments on two Chinese medical multiple-choice question datasets and one Chinese open-domain medical Q&A dataset with two LLM turbos demonstrate the superiority of HyKGE in terms of accuracy and explainability.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. Retrieval-based language models and applications. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 6: Tutorial Abstracts). 41–46.
  2. Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection. arXiv preprint arXiv:2310.11511 (2023).
  3. Hiteshwar Kumar Azad and Akshay Deepak. 2019. Query expansion techniques for information retrieval: A survey. Information Processing & Management (Sept. 2019), 1698–1735.
  4. Graph of Thoughts: Solving Elaborate Problems with Large Language Models. arXiv:2308.09687 [cs.CL]
  5. Factual Error Correction for Abstractive Summarization Models. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, Online, 6251–6258. https://doi.org/10.18653/v1/2020.emnlp-main.506
  6. Reading Wikipedia to Answer Open-Domain Questions. arXiv:1704.00051 [cs.CL]
  7. UnitedQA: A Hybrid Approach for Open Domain Question Answering. arXiv:2101.00178 [cs.CL]
  8. Complexity-Based Prompting for Multi-Step Reasoning. arXiv:2210.00720 [cs.CL]
  9. Precise Zero-Shot Dense Retrieval without Relevance Labels. arXiv:2212.10496 [cs.IR]
  10. Rethinking with Retrieval: Faithful Large Language Model Inference. arXiv:2301.00303 [cs.CL]
  11. Building a Pediatric Medical Corpus: Word Segmentation and Named Entity Annotation. In Workshop on Chinese Lexical Semantics. 652–664.
  12. Atlas: Few-shot Learning with Retrieval Augmented Language Models. arXiv:2208.03299 [cs.CL]
  13. Survey of hallucination in natural language generation. Comput. Surveys 55, 12 (2023), 1–38.
  14. Survey of Hallucination in Natural Language Generation. ACM Comput. Surv. 55, 12, Article 248 (mar 2023), 38 pages. https://doi.org/10.1145/3571730
  15. Large language models struggle to learn long-tail knowledge. In International Conference on Machine Learning. PMLR, 15696–15707.
  16. KALA: Knowledge-Augmented Language Model Adaptation. arXiv:2204.10555 [cs.CL]
  17. Dense Passage Retrieval for Open-Domain Question Answering. arXiv:2004.04906 [cs.CL]
  18. Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In ICLR.
  19. Large Language Models are Zero-Shot Reasoners. arXiv:2205.11916 [cs.CL]
  20. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. arXiv:2005.11401 [cs.CL]
  21. Unified Named Entity Recognition as Word-Word Relation Classification. arXiv:2112.10070 [cs.CL]
  22. Chain-of-Knowledge: Grounding Large Language Models via Dynamic Knowledge Adapting over Heterogeneous Sources. arXiv:2305.13269 [cs.CL]
  23. ChatDoctor: A Medical Chat Model Fine-Tuned on a Large Language Model Meta-AI (LLaMA) Using Medical Domain Knowledge. arXiv:2303.14070 [cs.CL]
  24. Towards General Text Embeddings with Multi-stage Contrastive Learning. arXiv:2308.03281 [cs.CL]
  25. Deductive Verification of Chain-of-Thought Reasoning. arXiv:2306.03872 [cs.CL]
  26. Reasoning on Graphs: Faithful and Interpretable Large Language Model Reasoning. arXiv:2310.01061 [cs.CL]
  27. Chain-of-Skills: A Configurable Model for Open-domain Question Answering. arXiv:2305.03130 [cs.CL]
  28. OpenAI. 2022. Introducing ChatGPT. https://openai.com/blog/chatgpt.
  29. OpenAI. 2023. GPT-4 Technical Report. ArXiv abs/2303.08774 (2023).
  30. RocketQA: An Optimized Training Approach to Dense Passage Retrieval for Open-Domain Question Answering. arXiv:2010.08191 [cs.CL]
  31. IEEE (2020).
  32. Think-on-Graph: Deep and Responsible Reasoning of Large Language Model on Knowledge Graph. arXiv:2307.07697 [cs.CL]
  33. SPARQA: Skeleton-based Semantic Parsing for Complex Questions over Knowledge Bases. arXiv:2003.13956 [cs.CL]
  34. Knowledge-Driven CoT: Exploring Faithful Reasoning in LLMs for Knowledge-intensive Question Answering. arXiv:2308.13259 [cs.CL]
  35. CMB: A Comprehensive Medical Benchmark in Chinese. arXiv:2308.08833 [cs.CL]
  36. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. arXiv:2201.11903 [cs.CL]
  37. C-Pack: Packaged Resources To Advance General Chinese Embedding. arXiv:2309.07597 [cs.CL]
  38. Baichuan 2: Open Large-scale Language Models. arXiv:2309.10305 [cs.CL]
  39. Large Language Models as Optimizers. arXiv:2309.03409 [cs.LG]
  40. Tree of Thoughts: Deliberate Problem Solving with Large Language Models. arXiv:2305.10601 [cs.CL]
  41. KG-FiD: Infusing Knowledge Graph in Fusion-in-Decoder for Open-Domain Question Answering. arXiv:2110.04330 [cs.CL]
  42. Chain-of-Note: Enhancing Robustness in Retrieval-Augmented Language Models. arXiv:2311.09210 [cs.CL]
  43. Hui Zeng. 2023. Measuring Massive Multitask Chinese Understanding. arXiv:2304.12986 [cs.CL]
  44. HuatuoGPT, towards Taming Language Model to Be a Doctor. arXiv:2305.15075 [cs.CL]
  45. CBLUE: A Chinese Biomedical Language Understanding Evaluation Benchmark. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Association for Computational Linguistics, Dublin, Ireland, 7888–7915. https://aclanthology.org/2022.acl-long.544
  46. Retrieving and Reading: A Comprehensive Survey on Open-domain Question Answering. arXiv:2101.00774 [cs.AI]
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Xinke Jiang (16 papers)
  2. Ruizhe Zhang (46 papers)
  3. Yongxin Xu (15 papers)
  4. Rihong Qiu (4 papers)
  5. Yue Fang (16 papers)
  6. Zhiyuan Wang (102 papers)
  7. Jinyi Tang (1 paper)
  8. Hongxin Ding (3 papers)
  9. Xu Chu (66 papers)
  10. Junfeng Zhao (22 papers)
  11. Yasha Wang (47 papers)
Citations (12)