Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ConceptNet infused DialoGPT for Underlying Commonsense Understanding and Reasoning in Dialogue Response Generation (2209.15109v1)

Published 29 Sep 2022 in cs.CL

Abstract: The pre-trained conversational models still fail to capture the implicit commonsense (CS) knowledge hidden in the dialogue interaction, even though they were pre-trained with an enormous dataset. In order to build a dialogue agent with CS capability, we firstly inject external knowledge into a pre-trained conversational model to establish basic commonsense through efficient Adapter tuning (Section 4). Secondly, we propose the ``two-way learning'' method to enable the bidirectional relationship between CS knowledge and sentence pairs so that the model can generate a sentence given the CS triplets, also generate the underlying CS knowledge given a sentence (Section 5). Finally, we leverage this integrated CS capability to improve open-domain dialogue response generation so that the dialogue agent is capable of understanding the CS knowledge hidden in dialogue history on top of inferring related other knowledge to further guide response generation (Section 6). The experiment results demonstrate that CS_Adapter fusion helps DialoGPT to be able to generate series of CS knowledge. And the DialoGPT+CS_Adapter response model adapted from CommonGen training can generate underlying CS triplets that fits better to dialogue context.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Ye Liu (153 papers)
  2. Wolfgang Maier (7 papers)
  3. Wolfgang Minker (18 papers)
  4. Stefan Ultes (32 papers)
Citations (2)