Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning to Retrieve Entity-Aware Knowledge and Generate Responses with Copy Mechanism for Task-Oriented Dialogue Systems (2012.11937v1)

Published 22 Dec 2020 in cs.CL, cs.AI, and cs.LG

Abstract: Task-oriented conversational modeling with unstructured knowledge access, as track 1 of the 9th Dialogue System Technology Challenges (DSTC 9), requests to build a system to generate response given dialogue history and knowledge access. This challenge can be separated into three subtasks, (1) knowledge-seeking turn detection, (2) knowledge selection, and (3) knowledge-grounded response generation. We use pre-trained LLMs, ELECTRA and RoBERTa, as our base encoder for different subtasks. For subtask 1 and 2, the coarse-grained information like domain and entity are used to enhance knowledge usage. For subtask 3, we use a latent variable to encode dialog history and selected knowledge better and generate responses combined with copy mechanism. Meanwhile, some useful post-processing strategies are performed on the model's final output to make further knowledge usage in the generation task. As shown in released evaluation results, our proposed system ranks second under objective metrics and ranks fourth under human metrics.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Chao-Hong Tan (9 papers)
  2. Xiaoyu Yang (85 papers)
  3. Zi'ou Zheng (4 papers)
  4. Tianda Li (10 papers)
  5. Yufei Feng (18 papers)
  6. Jia-Chen Gu (42 papers)
  7. Quan Liu (116 papers)
  8. Dan Liu (74 papers)
  9. Zhen-Hua Ling (114 papers)
  10. Xiaodan Zhu (94 papers)
Citations (9)