Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dual Task Framework for Improving Persona-grounded Dialogue Dataset (2202.05435v2)

Published 11 Feb 2022 in cs.CL, cs.AI, and cs.LG

Abstract: This paper introduces a simple yet effective data-centric approach for the task of improving persona-conditioned dialogue agents. Prior model-centric approaches unquestioningly depend on the raw crowdsourced benchmark datasets such as Persona-Chat. In contrast, we aim to fix annotation artifacts in benchmarking, which is orthogonally applicable to any dialogue model. Specifically, we augment relevant personas to improve dialogue dataset/agent, by leveraging the primal-dual structure of the two tasks, predicting dialogue responses and personas based on each other. Experiments on Persona-Chat show that our approach outperforms pre-trained LMs by an 11.7 point gain in terms of accuracy.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Minju Kim (12 papers)
  2. Beong-woo Kwak (12 papers)
  3. Youngwook Kim (30 papers)
  4. Hong-in Lee (2 papers)
  5. Seung-won Hwang (59 papers)
  6. Jinyoung Yeo (46 papers)
Citations (7)