Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Focus-Constrained Attention Mechanism for CVAE-based Response Generation (2009.12102v1)

Published 25 Sep 2020 in cs.CL

Abstract: To model diverse responses for a given post, one promising way is to introduce a latent variable into Seq2Seq models. The latent variable is supposed to capture the discourse-level information and encourage the informativeness of target responses. However, such discourse-level information is often too coarse for the decoder to be utilized. To tackle it, our idea is to transform the coarse-grained discourse-level information into fine-grained word-level information. Specifically, we firstly measure the semantic concentration of corresponding target response on the post words by introducing a fine-grained focus signal. Then, we propose a focus-constrained attention mechanism to take full advantage of focus in well aligning the input to the target response. The experimental results demonstrate that by exploiting the fine-grained signal, our model can generate more diverse and informative responses compared with several state-of-the-art models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zhi Cui (5 papers)
  2. Yanran Li (32 papers)
  3. Jiayi Zhang (160 papers)
  4. Jianwei Cui (18 papers)
  5. Chen Wei (72 papers)
  6. Bin Wang (751 papers)
Citations (7)

Summary

We haven't generated a summary for this paper yet.