Prompting Explicit and Implicit Knowledge for Multi-hop Question Answering Based on Human Reading Process (2402.19350v6)
Abstract: Pre-trained LLMs (PLMs) leverage chains-of-thought (CoT) to simulate human reasoning and inference processes, achieving proficient performance in multi-hop QA. However, a gap persists between PLMs' reasoning abilities and those of humans when tackling complex problems. Psychological studies suggest a vital connection between explicit information in passages and human prior knowledge during reading. Nevertheless, current research has given insufficient attention to linking input passages and PLMs' pre-training-based knowledge from the perspective of human cognition studies. In this study, we introduce a Prompting Explicit and Implicit knowledge (PEI) framework, which uses prompts to connect explicit and implicit knowledge, aligning with human reading process for multi-hop QA. We consider the input passages as explicit knowledge, employing them to elicit implicit knowledge through unified prompt reasoning. Furthermore, our model incorporates type-specific reasoning via prompts, a form of implicit knowledge. Experimental results show that PEI performs comparably to the state-of-the-art on HotpotQA. Ablation studies confirm the efficacy of our model in bridging and integrating explicit and implicit knowledge.
- Noureldin Mohamed Abdelaal and Amal Saleh Sase. 2014. Relationship between prior knowledge and reading comprehension. Advances in Language and Literary Studies, 5(6):125–131.
- Beamqa: Multi-hop knowledge graph question answering with sequence-to-sequence prediction and beam search. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 781–790.
- Effects of topic interest and prior knowledge on reading comprehension. Reading research quarterly, pages 497–504.
- Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150.
- On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, pages 610–623.
- Rpa: reasoning path augmentation in iterative retrieving for multi-hop qa. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 12598–12606.
- A survey of chain of thought reasoning: Advances, frontiers and future. arXiv preprint arXiv:2309.15402.
- Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555.
- Mark A Clarke and Sandra Silberstein. 1977. Toward a realization of psycholinguistic principles in the esl reading class 1. Language learning, 27(1):135–154.
- Prompt-based conservation learning for multi-hop question answering. In Proceedings of the 29th International Conference on Computational Linguistics, pages 1791–1800.
- Hierarchical graph network for multi-hop question answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 8823–8838.
- Integration of word meaning and world knowledge in language comprehension. science, 304(5669):438–441.
- Constructing a multi-hop qa dataset for comprehensive evaluation of reasoning steps. In Proceedings of the 28th International Conference on Computational Linguistics, pages 6609–6625.
- LIDA: Lexical-based imbalanced data augmentation for content moderation. In Proceedings of the 37th Pacific Asia Conference on Language, Information and Computation, pages 59–69, Hong Kong, China. Association for Computational Linguistics.
- Charles Jin and Martin Rinard. 2023. Evidence of meaning in language models trained on programs. arXiv preprint arXiv:2305.11169.
- Few-shot reranking for multi-hop qa via language model prompting. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 15882–15897.
- Large language models are zero-shot reasoners. arXiv preprint arXiv:2205.11916.
- The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3045–3059.
- Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880.
- Asynchronous multi-grained graph network for interpretable multi-hop reading comprehension. In IJCAI, pages 3857–3863.
- Xiang Lisa Li and Percy Liang. 2021. Prefix-tuning: Optimizing continuous prompts for generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4582–4597.
- Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9):1–35.
- P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks. arXiv preprint arXiv:2110.07602.
- Ilya Loshchilov and Frank Hutter. 2018. Decoupled weight decay regularization. In International Conference on Learning Representations.
- Multi-hop reading comprehension through question decomposition and rescoring. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6097–6109.
- Answering while summarizing: Multi-task learning for multi-hop qa with evidence extraction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 2335–2345.
- Unsupervised question decomposition for question answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 8864–8880.
- Dynamically fused graph network for multi-hop reasoning. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6140–6150.
- Squad: 100,000+ questions for machine comprehension of text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 2383–2392.
- Is graph structure necessary for multi-hop question answering? In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 7187–7192.
- Frank Smith. 1971. Understanding reading: A psycholinguistic analysis of reading and learning to read. Holt, Rinehart and Winston, New York.
- Do multi-hop question answering systems know how to answer the single-hop sub-questions? In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 3244–3249.
- Interleaving retrieval with chain-of-thought reasoning for knowledge-intensive multi-step questions. arXiv preprint arXiv:2212.10509.
- Musique: Multihop questions via single-hop question composition. Transactions of the Association for Computational Linguistics, 10:539–554.
- Select, answer and explain: Interpretable multi-hop reading comprehension over multiple documents. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 9073–9080.
- Iteratively prompt pre-trained language models for chain of thought. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 2714–2730.
- Chain of thought prompting elicits reasoning in large language models. arXiv preprint arXiv:2201.11903.
- Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations, pages 38–45.
- Graph-free multi-hop reading comprehension: A select-to-guide strategy. arXiv preprint arXiv:2107.11823.
- Hotpotqa: A dataset for diverse, explainable multi-hop question answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2369–2380.
- Koh Moy Yin. 1985. The role of prior knowledge in reading comprehension.
- Beam retrieval: General end-to-end retrieval for multi-hop question answering. arXiv preprint arXiv:2308.08973.
- Automatic chain of thought prompting in large language models. arXiv preprint arXiv:2210.03493.
- Guangming Huang (22 papers)
- Yunfei Long (26 papers)
- Cunjin Luo (2 papers)
- Jiaxing Shen (14 papers)
- Xia Sun (6 papers)