Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Deep Active Learning for Dialogue Generation (1612.03929v5)

Published 12 Dec 2016 in cs.CL, cs.AI, and cs.NE

Abstract: We propose an online, end-to-end, neural generative conversational model for open-domain dialogue. It is trained using a unique combination of offline two-phase supervised learning and online human-in-the-loop active learning. While most existing research proposes offline supervision or hand-crafted reward functions for online reinforcement, we devise a novel interactive learning mechanism based on hamming-diverse beam search for response generation and one-character user-feedback at each step. Experiments show that our model inherently promotes the generation of semantically relevant and interesting responses, and can be used to train agents with customized personas, moods and conversational styles.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Nabiha Asghar (9 papers)
  2. Pascal Poupart (80 papers)
  3. Xin Jiang (242 papers)
  4. Hang Li (277 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.