Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Enhancing Pre-trained Models with Text Structure Knowledge for Question Generation (2209.04179v1)

Published 9 Sep 2022 in cs.CL and cs.AI

Abstract: Today the pre-trained LLMs achieve great success for question generation (QG) task and significantly outperform traditional sequence-to-sequence approaches. However, the pre-trained models treat the input passage as a flat sequence and are thus not aware of the text structure of input passage. For QG task, we model text structure as answer position and syntactic dependency, and propose answer localness modeling and syntactic mask attention to address these limitations. Specially, we present localness modeling with a Gaussian bias to enable the model to focus on answer-surrounded context, and propose a mask attention mechanism to make the syntactic structure of input passage accessible in question generation process. Experiments on SQuAD dataset show that our proposed two modules improve performance over the strong pre-trained model ProphetNet, and combing them together achieves very competitive results with the state-of-the-art pre-trained model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Zichen Wu (3 papers)
  2. Xin Jia (6 papers)
  3. Fanyi Qu (7 papers)
  4. Yunfang Wu (50 papers)
Citations (4)

Summary

We haven't generated a summary for this paper yet.