2000 character limit reached
Retrieval Enhanced Model for Commonsense Generation (2105.11174v1)
Published 24 May 2021 in cs.CL and cs.AI
Abstract: Commonsense generation is a challenging task of generating a plausible sentence describing an everyday scenario using provided concepts. Its requirement of reasoning over commonsense knowledge and compositional generalization ability even puzzles strong pre-trained language generation models. We propose a novel framework using retrieval methods to enhance both the pre-training and fine-tuning for commonsense generation. We retrieve prototype sentence candidates by concept matching and use them as auxiliary input. For fine-tuning, we further boost its performance with a trainable sentence retriever. We demonstrate experimentally on the large-scale CommonGen benchmark that our approach achieves new state-of-the-art results.
- Han Wang (418 papers)
- Yang Liu (2253 papers)
- Chenguang Zhu (100 papers)
- Linjun Shou (53 papers)
- Ming Gong (246 papers)
- Yichong Xu (42 papers)
- Michael Zeng (76 papers)