Graph Guided Question Answer Generation for Procedural Question-Answering (2401.13594v1)
Abstract: In this paper, we focus on task-specific question answering (QA). To this end, we introduce a method for generating exhaustive and high-quality training data, which allows us to train compact (e.g., run on a mobile device), task-specific QA models that are competitive against GPT variants. The key technological enabler is a novel mechanism for automatic question-answer generation from procedural text which can ingest large amounts of textual instructions and produce exhaustive in-domain QA training data. While current QA data generation methods can produce well-formed and varied data, their non-exhaustive nature is sub-optimal for training a QA model. In contrast, we leverage the highly structured aspect of procedural text and represent each step and the overall flow of the procedure as graphs. We then condition on graph nodes to automatically generate QA pairs in an exhaustive and controllable manner. Comprehensive evaluations of our method show that: 1) small models trained with our data achieve excellent performance on the target QA task, even exceeding that of GPT3 and ChatGPT despite being several orders of magnitude smaller. 2) semantic coverage is the key indicator for downstream QA performance. Crucially, while LLMs excel at syntactic diversity, this does not necessarily result in improvements on the end QA model. In contrast, the higher semantic coverage provided by our method is critical for QA performance.
- Synthetic QA corpora generation with roundtrip consistency. In Proc. of the Annual Meeting of the Association for Computational Linguistics.
- Abstract meaning representation for sembanking. In Proc. of the Linguistic Annotation Workshop and Interoperability with Discourse.
- BBC. BBC good food. https://www.bbcgoodfood.com/about-bbc-good-food.
- Language models are few-shot learners. In Conference on Neural Information Processing systems.
- Learning to Ask More: Semi-Autoregressive Sequential Question Generation under Dual-Graph Interaction. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 225–237, Online. Association for Computational Linguistics.
- Ying-Hong Chan and Yao-Chung Fan. 2019. A recurrent BERT-based model for question generation. In Workshop on Machine Reading for Question Answering.
- Aligning Actions Across Recipe Graphs. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 6930–6942, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Searchqa: A new Q&A dataset augmented with context from a search engine. ArXiv, abs/1704.05179.
- Flow graph to video grounding for weakly-supervised multi-step localization. In Computer Vision – ECCV 2022, pages 319–335, Cham. Springer Nature Switzerland.
- Template-Based Question Generation from Retrieved Sentences for Improved Unsupervised Question Answering. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 4508–4513, Online. Association for Computational Linguistics.
- Two-Stage Synthesis Networks for Transfer Learning in Machine Comprehension. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 835–844, Copenhagen, Denmark. Association for Computational Linguistics.
- DiffuSeq: Sequence to sequence text generation with diffusion models. In International Conference on Learning Representations.
- Structural analysis of cooking preparation steps in japanese. In Proceedings of the Fifth International Workshop on on Information Retrieval with Asian Languages, IRAL ’00, page 157–164, New York, NY, USA. Association for Computing Machinery.
- DuReader: a Chinese machine reading comprehension dataset from real-world applications. In Proceedings of the Workshop on Machine Reading for Question Answering.
- Michael Heilman and Noah A. Smith. 2010. Good Question! Statistical Ranking for Question Generation. In Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, pages 609–617, Los Angeles, California. Association for Computational Linguistics.
- Brad Jacob. 2020. AMRLib: A python library that makes AMR parsing, generation and visualization simple. https://github.com/bjascob/amrlib.
- TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1601–1611, Vancouver, Canada. Association for Computational Linguistics.
- A Diversity-Promoting Objective Function for Neural Conversation Models. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 110–119, San Diego, California. Association for Computational Linguistics.
- Yoshio Momouchi. 1980. Control structures for actions in procedural texts and PT-chart. In Conference on Computational Linguistics.
- DocAMR: Multi-sentence AMR representation and evaluation.
- MS MARCO: A human generated machine reading comprehension dataset. In Conference on Neural Information Processing Systems - Workshops.
- Suraf Patil. 2020. Question generation using transformers. https://github.com/patil-suraj/question_generation.
- Designing multimodal datasets for nlp challenges. arXiv:2105.05999.
- Designing multimodal datasets for NLP challenges. arXiv:2105.05999.
- Asking It All: Generating Contextualized Questions for any Semantic Role. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 1429–1441, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21(140):1–67.
- SQuAD: 100,000+ Questions for Machine Comprehension of Text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 2383–2392, Austin, Texas. Association for Computational Linguistics.
- Geetanjali Rakshit and Jeffrey Flanigan. 2021. ASQ: automatically generating question-answer pairs using AMRs. arXiv:2105.10023.
- CoQA: A conversational question answering challenge. Transactions of the Association for Computational Linguistics, 7:249–266.
- The first question generation shared task evaluation challenge. In International Natural Language Generation Conference.
- Self-Attention Architectures for Answer-Agnostic Neural Question Generation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6027–6032, Florence, Italy. Association for Computational Linguistics.
- BLEURT: Learning Robust Metrics for Text Generation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7881–7892, Online. Association for Computational Linguistics.
- A unified query-based generative model for question generation and question answering. arXiv:1709.01058.
- Stanford Alpaca: An instruction-following LLaMA model. https://github.com/tatsu-lab/stanford_alpaca.
- Neural question generation with answer pivot. In AAAI Conference on Artificial Intelligence.
- Want to reduce labeling cost? GPT-3 can help. In Conference on Empirical Methods in Natural Language Processing.
- On decoding strategies for neural text generators. TACL.
- RecipeQA: A challenge dataset for multimodal comprehension of cooking recipes. In Conference on Empirical Methods in Natural Language Processing.
- English Recipe Flow Graph Corpus. In Proceedings of the Twelfth Language Resources and Evaluation Conference, pages 5187–5194, Marseille, France. European Language Resources Association.
- Just ask: Learning to answer questions from millions of narrated videos. In International Conference on Computer Vision.
- HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2369–2380, Brussels, Belgium. Association for Computational Linguistics.
- Selecting better samples from pre-trained LLMs: A case study on question generation. arXiv:2209.11000.
- Paragraph-level Neural Question Generation with Maxout Pointer and Gated Self-attention Networks. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 3901–3910, Brussels, Belgium. Association for Computational Linguistics.
- Neural question generation from text: A preliminary study. In Natural Language Processing and Chinese Computing.
- Hai X. Pham (6 papers)
- Isma Hadji (14 papers)
- Xinnuo Xu (12 papers)
- Ziedune Degutyte (1 paper)
- Jay Rainey (1 paper)
- Evangelos Kazakos (13 papers)
- Afsaneh Fazly (9 papers)
- Georgios Tzimiropoulos (86 papers)
- Brais Martinez (38 papers)