Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sub-domain Modelling for Dialogue Management with Hierarchical Reinforcement Learning (1706.06210v2)

Published 19 Jun 2017 in cs.CL and cs.AI

Abstract: Human conversation is inherently complex, often spanning many different topics/domains. This makes policy learning for dialogue systems very challenging. Standard flat reinforcement learning methods do not provide an efficient framework for modelling such dialogues. In this paper, we focus on the under-explored problem of multi-domain dialogue management. First, we propose a new method for hierarchical reinforcement learning using the option framework. Next, we show that the proposed architecture learns faster and arrives at a better policy than the existing flat ones do. Moreover, we show how pretrained policies can be adapted to more complex systems with an additional set of new actions. In doing that, we show that our approach has the potential to facilitate policy optimisation for more sophisticated multi-domain dialogue systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Paweł Budzianowski (27 papers)
  2. Stefan Ultes (32 papers)
  3. Pei-Hao Su (25 papers)
  4. Nikola Mrkšić (30 papers)
  5. Tsung-Hsien Wen (27 papers)
  6. Iñigo Casanueva (18 papers)
  7. Lina Rojas-Barahona (11 papers)
  8. Milica Gašić (57 papers)
Citations (49)

Summary

We haven't generated a summary for this paper yet.