Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Retrieval Enhanced Model for Commonsense Generation (2105.11174v1)

Published 24 May 2021 in cs.CL and cs.AI

Abstract: Commonsense generation is a challenging task of generating a plausible sentence describing an everyday scenario using provided concepts. Its requirement of reasoning over commonsense knowledge and compositional generalization ability even puzzles strong pre-trained language generation models. We propose a novel framework using retrieval methods to enhance both the pre-training and fine-tuning for commonsense generation. We retrieve prototype sentence candidates by concept matching and use them as auxiliary input. For fine-tuning, we further boost its performance with a trainable sentence retriever. We demonstrate experimentally on the large-scale CommonGen benchmark that our approach achieves new state-of-the-art results.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Han Wang (418 papers)
  2. Yang Liu (2253 papers)
  3. Chenguang Zhu (100 papers)
  4. Linjun Shou (53 papers)
  5. Ming Gong (246 papers)
  6. Yichong Xu (42 papers)
  7. Michael Zeng (76 papers)
Citations (29)