Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Probing Neural Dialog Models for Conversational Understanding (2006.08331v1)

Published 7 Jun 2020 in cs.CL, cs.AI, cs.LG, and stat.ML

Abstract: The predominant approach to open-domain dialog generation relies on end-to-end training of neural models on chat datasets. However, this approach provides little insight as to what these models learn (or do not learn) about engaging in dialog. In this study, we analyze the internal representations learned by neural open-domain dialog systems and evaluate the quality of these representations for learning basic conversational skills. Our results suggest that standard open-domain dialog systems struggle with answering questions, inferring contradiction, and determining the topic of conversation, among other tasks. We also find that the dyadic, turn-taking nature of dialog is not fully leveraged by these models. By exploring these limitations, we highlight the need for additional research into architectures and training methods that can better capture high-level information about dialog.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Abdelrhman Saleh (4 papers)
  2. Tovly Deutsch (2 papers)
  3. Stephen Casper (40 papers)
  4. Yonatan Belinkov (111 papers)
  5. Stuart Shieber (6 papers)
Citations (13)