2000 character limit reached
Dual Task Framework for Improving Persona-grounded Dialogue Dataset (2202.05435v2)
Published 11 Feb 2022 in cs.CL, cs.AI, and cs.LG
Abstract: This paper introduces a simple yet effective data-centric approach for the task of improving persona-conditioned dialogue agents. Prior model-centric approaches unquestioningly depend on the raw crowdsourced benchmark datasets such as Persona-Chat. In contrast, we aim to fix annotation artifacts in benchmarking, which is orthogonally applicable to any dialogue model. Specifically, we augment relevant personas to improve dialogue dataset/agent, by leveraging the primal-dual structure of the two tasks, predicting dialogue responses and personas based on each other. Experiments on Persona-Chat show that our approach outperforms pre-trained LMs by an 11.7 point gain in terms of accuracy.
- Minju Kim (12 papers)
- Beong-woo Kwak (12 papers)
- Youngwook Kim (30 papers)
- Hong-in Lee (2 papers)
- Seung-won Hwang (59 papers)
- Jinyoung Yeo (46 papers)