FOLLOWUPQG: Towards Information-Seeking Follow-up Question Generation (2309.05007v2)
Abstract: Humans ask follow-up questions driven by curiosity, which reflects a creative human cognitive process. We introduce the task of real-world information-seeking follow-up question generation (FQG), which aims to generate follow-up questions seeking a more in-depth understanding of an initial question and answer. We construct FOLLOWUPQG, a dataset of over 3K real-world (initial question, answer, follow-up question) tuples collected from a Reddit forum providing layman-friendly explanations for open-ended questions. In contrast to existing datasets, questions in FOLLOWUPQG use more diverse pragmatic strategies to seek information, and they also show higher-order cognitive skills (such as applying and relating). We evaluate current question generation models on their efficacy for generating follow-up questions, exploring how to generate specific types of follow-up questions based on step-by-step demonstrations. Our results validate FOLLOWUPQG as a challenging benchmark, as model-generated questions are adequate but far from human-raised questions in terms of informativeness and complexity.
- A taxonomy for learning, teaching, and assessing: A revision of bloom’s taxonomy of educational objectives, abridged edition. White Plains, NY: Longman.
- GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow.
- Learningq: A large-scale dataset for educational question generation. In International Conference on Web and Social Media (ICWSM), pages 481–490.
- JT Dillon. 1988. Questioning and teaching. a manual of practice.
- Xinya Du and Claire Cardie. 2018. Harvesting paragraph-level question-answer pairs from wikipedia. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 1907–1917.
- Learning to ask: Neural question generation for reading comprehension. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 1342–1352.
- Question generation for question answering. In Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 866–874.
- ELI5: long form question answering. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 3558–3567.
- Question generation for reading comprehension assessment by modeling how and what to ask. In Findings of Annual Meeting of the Association for Computational Linguistics (ACL), pages 2131–2146.
- Arthur C Graesser and Natalie K Person. 1994. Question asking during tutoring. American educational research journal, 31(1):104–137.
- Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings.
- Inquisitive question generation for high level text comprehension. In Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6544–6555.
- Vaibhav Kumar and Alan W. Black. 2020. Clarq: A large-scale and diverse dataset for clarification question generation. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 7296–7301.
- Quiz design task: Helping teachers create quizzes with automated question generation. CoRR, abs/2205.01730.
- Alon Lavie and Abhaya Agarwal. 2007. METEOR: An automatic metric for MT evaluation with high levels of correlation with human judgments. In Proceedings of the Second Workshop on Statistical Machine Translation, pages 228–231.
- BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 7871–7880.
- Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74–81.
- Learning to generate questions by learning what not to generate. In International World Wide Web Conference (WWW), pages 1106–1118.
- Ask what’s missing and what’s useful: Improving clarification question generation using global knowledge. In Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL-HLT), pages 4300–4312.
- Christopher Malon and Bing Bai. 2020. Generating followup questions for interpretable multi-hop question answering. ArXiv, abs/2002.12344.
- Naomi Miyake and Donald A Norman. 1979. To ask a question, one must know enough to know what is not known. Journal of verbal learning and verbal behavior, 18(3):357–364.
- Mixqg: Neural question generation with mixed answer types. CoRR, abs/2110.08175.
- Let’s ask again: Refine network for automatic question generation. In Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3312–3321.
- Unsupervised multi-hop question answering by question generation. In Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL-HLT), pages 5866–5880.
- Zero-shot fact verification by claim generation. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 476–483.
- Recent advances in neural question generation.
- Semantic graphs for generating deep questions. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 1463–1475.
- Bleu: a method for automatic evaluation of machine translation. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 311–318. ACL.
- Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21:140:1–140:67.
- Know what you don’t know: Unanswerable questions for squad. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 784–789.
- Squad: 100, 000+ questions for machine comprehension of text. In Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 2383–2392.
- Ashwin Ram. 1991. A theory of questions and question asking. Journal of the Learning Sciences, 1(3-4):273–318.
- Sudha Rao and Hal Daumé III. 2018. Learning to ask good questions: Ranking clarification questions using neural expected value of perfect information. In Annual Meeting of the Association for Computational Linguistics (ACL), pages 2737–2746.
- Sudha Rao and Hal Daumé III. 2019. Answer-based adversarial training for generating clarification questions. In Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL-HLT), pages 143–155.
- CoQA: A conversational question answering challenge. Transactions of the Association for Computational Linguistics, 7:249–266.
- Learning to retrieve engaging follow-up queries. In Findings of the Association for Computational Linguistics: EACL 2023, pages 2009–2016, Dubrovnik, Croatia. Association for Computational Linguistics.
- Summarization programs: Interpretable abstractive summarization with neural modular trees.
- Neural models for key phrase extraction and question generation. In Workshop on Machine Reading for Question Answering@ACL, pages 78–88.
- A multi-agent communication framework for question-worthy phrase extraction and question generation. In AAAI Conference on Artificial Intelligence (AAAI), pages 7168–7175.
- Chain of thought prompting elicits reasoning in large language models. ArXiv, abs/2201.11903.
- Hotpotqa: A dataset for diverse, explainable multi-hop question answering. In Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 2369–2380.
- Paragraph-level neural question generation with maxout pointer and gated self-attention networks. In Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3901–3910.