Towards Coherent and Cohesive Long-form Text Generation (1811.00511v2)
Abstract: Generating coherent and cohesive long-form texts is a challenging task. Previous works relied on large amounts of human-generated texts to train neural LLMs. However, few attempted to explicitly improve neural LLMs from the perspectives of coherence and cohesion. In this work, we propose a new neural LLM that is equipped with two neural discriminators which provide feedback signals at the levels of sentence (cohesion) and paragraph (coherence). Our model is trained using a simple yet efficient variant of policy gradient, called negative-critical sequence training, which is proposed to eliminate the need of training a separate critic for estimating baseline. Results demonstrate the effectiveness of our approach, showing improvements over the strong baseline -- recurrent attention-based bidirectional MLE-trained neural LLM.
- Woon Sang Cho (4 papers)
- Pengchuan Zhang (58 papers)
- Yizhe Zhang (127 papers)
- Xiujun Li (37 papers)
- Michel Galley (50 papers)
- Chris Brockett (37 papers)
- Mengdi Wang (199 papers)
- Jianfeng Gao (344 papers)