Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Dual-Attention Hierarchical Recurrent Neural Network for Dialogue Act Classification (1810.09154v3)

Published 22 Oct 2018 in cs.CL

Abstract: Recognising dialogue acts (DA) is important for many natural language processing tasks such as dialogue generation and intention recognition. In this paper, we propose a dual-attention hierarchical recurrent neural network for DA classification. Our model is partially inspired by the observation that conversational utterances are normally associated with both a DA and a topic, where the former captures the social act and the latter describes the subject matter. However, such a dependency between DAs and topics has not been utilised by most existing systems for DA classification. With a novel dual task-specific attention mechanism, our model is able, for utterances, to capture information about both DAs and topics, as well as information about the interactions between them. Experimental results show that by modelling topic as an auxiliary task, our model can significantly improve DA classification, yielding better or comparable performance to the state-of-the-art method on three public datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Ruizhe Li (40 papers)
  2. Chenghua Lin (127 papers)
  3. Matthew Collinson (3 papers)
  4. Xiao Li (354 papers)
  5. Guanyi Chen (26 papers)
Citations (49)