Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Prompting Explicit and Implicit Knowledge for Multi-hop Question Answering Based on Human Reading Process (2402.19350v6)

Published 29 Feb 2024 in cs.CL

Abstract: Pre-trained LLMs (PLMs) leverage chains-of-thought (CoT) to simulate human reasoning and inference processes, achieving proficient performance in multi-hop QA. However, a gap persists between PLMs' reasoning abilities and those of humans when tackling complex problems. Psychological studies suggest a vital connection between explicit information in passages and human prior knowledge during reading. Nevertheless, current research has given insufficient attention to linking input passages and PLMs' pre-training-based knowledge from the perspective of human cognition studies. In this study, we introduce a Prompting Explicit and Implicit knowledge (PEI) framework, which uses prompts to connect explicit and implicit knowledge, aligning with human reading process for multi-hop QA. We consider the input passages as explicit knowledge, employing them to elicit implicit knowledge through unified prompt reasoning. Furthermore, our model incorporates type-specific reasoning via prompts, a form of implicit knowledge. Experimental results show that PEI performs comparably to the state-of-the-art on HotpotQA. Ablation studies confirm the efficacy of our model in bridging and integrating explicit and implicit knowledge.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. Noureldin Mohamed Abdelaal and Amal Saleh Sase. 2014. Relationship between prior knowledge and reading comprehension. Advances in Language and Literary Studies, 5(6):125–131.
  2. Beamqa: Multi-hop knowledge graph question answering with sequence-to-sequence prediction and beam search. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 781–790.
  3. Effects of topic interest and prior knowledge on reading comprehension. Reading research quarterly, pages 497–504.
  4. Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150.
  5. On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, pages 610–623.
  6. Rpa: reasoning path augmentation in iterative retrieving for multi-hop qa. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 12598–12606.
  7. A survey of chain of thought reasoning: Advances, frontiers and future. arXiv preprint arXiv:2309.15402.
  8. Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555.
  9. Mark A Clarke and Sandra Silberstein. 1977. Toward a realization of psycholinguistic principles in the esl reading class 1. Language learning, 27(1):135–154.
  10. Prompt-based conservation learning for multi-hop question answering. In Proceedings of the 29th International Conference on Computational Linguistics, pages 1791–1800.
  11. Hierarchical graph network for multi-hop question answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 8823–8838.
  12. Integration of word meaning and world knowledge in language comprehension. science, 304(5669):438–441.
  13. Constructing a multi-hop qa dataset for comprehensive evaluation of reasoning steps. In Proceedings of the 28th International Conference on Computational Linguistics, pages 6609–6625.
  14. LIDA: Lexical-based imbalanced data augmentation for content moderation. In Proceedings of the 37th Pacific Asia Conference on Language, Information and Computation, pages 59–69, Hong Kong, China. Association for Computational Linguistics.
  15. Charles Jin and Martin Rinard. 2023. Evidence of meaning in language models trained on programs. arXiv preprint arXiv:2305.11169.
  16. Few-shot reranking for multi-hop qa via language model prompting. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 15882–15897.
  17. Large language models are zero-shot reasoners. arXiv preprint arXiv:2205.11916.
  18. The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3045–3059.
  19. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880.
  20. Asynchronous multi-grained graph network for interpretable multi-hop reading comprehension. In IJCAI, pages 3857–3863.
  21. Xiang Lisa Li and Percy Liang. 2021. Prefix-tuning: Optimizing continuous prompts for generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4582–4597.
  22. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9):1–35.
  23. P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks. arXiv preprint arXiv:2110.07602.
  24. Ilya Loshchilov and Frank Hutter. 2018. Decoupled weight decay regularization. In International Conference on Learning Representations.
  25. Multi-hop reading comprehension through question decomposition and rescoring. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6097–6109.
  26. Answering while summarizing: Multi-task learning for multi-hop qa with evidence extraction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 2335–2345.
  27. Unsupervised question decomposition for question answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 8864–8880.
  28. Dynamically fused graph network for multi-hop reasoning. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6140–6150.
  29. Squad: 100,000+ questions for machine comprehension of text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 2383–2392.
  30. Is graph structure necessary for multi-hop question answering? In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 7187–7192.
  31. Frank Smith. 1971. Understanding reading: A psycholinguistic analysis of reading and learning to read. Holt, Rinehart and Winston, New York.
  32. Do multi-hop question answering systems know how to answer the single-hop sub-questions? In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 3244–3249.
  33. Interleaving retrieval with chain-of-thought reasoning for knowledge-intensive multi-step questions. arXiv preprint arXiv:2212.10509.
  34. Musique: Multihop questions via single-hop question composition. Transactions of the Association for Computational Linguistics, 10:539–554.
  35. Select, answer and explain: Interpretable multi-hop reading comprehension over multiple documents. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 9073–9080.
  36. Iteratively prompt pre-trained language models for chain of thought. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 2714–2730.
  37. Chain of thought prompting elicits reasoning in large language models. arXiv preprint arXiv:2201.11903.
  38. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations, pages 38–45.
  39. Graph-free multi-hop reading comprehension: A select-to-guide strategy. arXiv preprint arXiv:2107.11823.
  40. Hotpotqa: A dataset for diverse, explainable multi-hop question answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2369–2380.
  41. Koh Moy Yin. 1985. The role of prior knowledge in reading comprehension.
  42. Beam retrieval: General end-to-end retrieval for multi-hop question answering. arXiv preprint arXiv:2308.08973.
  43. Automatic chain of thought prompting in large language models. arXiv preprint arXiv:2210.03493.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Guangming Huang (22 papers)
  2. Yunfei Long (26 papers)
  3. Cunjin Luo (2 papers)
  4. Jiaxing Shen (14 papers)
  5. Xia Sun (6 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.