Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural Headline Generation with Sentence-wise Optimization (1604.01904v2)

Published 7 Apr 2016 in cs.CL

Abstract: Recently, neural models have been proposed for headline generation by learning to map documents to headlines with recurrent neural networks. Nevertheless, as traditional neural network utilizes maximum likelihood estimation for parameter optimization, it essentially constrains the expected training objective within word level rather than sentence level. Moreover, the performance of model prediction significantly relies on training data distribution. To overcome these drawbacks, we employ minimum risk training strategy in this paper, which directly optimizes model parameters in sentence level with respect to evaluation metrics and leads to significant improvements for headline generation. Experiment results show that our models outperforms state-of-the-art systems on both English and Chinese headline generation tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Ayana (2 papers)
  2. Shiqi Shen (14 papers)
  3. Yu Zhao (209 papers)
  4. Zhiyuan Liu (433 papers)
  5. Maosong Sun (337 papers)
Citations (55)

Summary

We haven't generated a summary for this paper yet.