Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Affective Neural Response Generation (1709.03968v1)

Published 12 Sep 2017 in cs.CL, cs.AI, cs.CY, cs.HC, and cs.IR

Abstract: Existing neural conversational models process natural language primarily on a lexico-syntactic level, thereby ignoring one of the most crucial components of human-to-human dialogue: its affective content. We take a step in this direction by proposing three novel ways to incorporate affective/emotional aspects into long short term memory (LSTM) encoder-decoder neural conversation models: (1) affective word embeddings, which are cognitively engineered, (2) affect-based objective functions that augment the standard cross-entropy loss, and (3) affectively diverse beam search for decoding. Experiments show that these techniques improve the open-domain conversational prowess of encoder-decoder networks by enabling them to produce emotionally rich responses that are more interesting and natural.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Nabiha Asghar (9 papers)
  2. Pascal Poupart (80 papers)
  3. Jesse Hoey (25 papers)
  4. Xin Jiang (242 papers)
  5. Lili Mou (79 papers)
Citations (152)