Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

BoB: BERT Over BERT for Training Persona-based Dialogue Models from Limited Personalized Data (2106.06169v2)

Published 11 Jun 2021 in cs.CL and cs.AI

Abstract: Maintaining consistent personas is essential for dialogue agents. Although tremendous advancements have been brought, the limited-scale of annotated persona-dense data are still barriers towards training robust and consistent persona-based dialogue models. In this work, we show how the challenges can be addressed by disentangling persona-based dialogue generation into two sub-tasks with a novel BERT-over-BERT (BoB) model. Specifically, the model consists of a BERT-based encoder and two BERT-based decoders, where one decoder is for response generation, and another is for consistency understanding. In particular, to learn the ability of consistency understanding from large-scale non-dialogue inference data, we train the second decoder in an unlikelihood manner. Under different limited data settings, both automatic and human evaluations demonstrate that the proposed model outperforms strong baselines in response quality and persona consistency.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Haoyu Song (21 papers)
  2. Yan Wang (733 papers)
  3. Kaiyan Zhang (33 papers)
  4. Wei-Nan Zhang (19 papers)
  5. Ting Liu (329 papers)
Citations (109)