Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural Latent Extractive Document Summarization (1808.07187v2)

Published 22 Aug 2018 in cs.CL, cs.AI, and cs.LG

Abstract: Extractive summarization models require sentence-level labels, which are usually created heuristically (e.g., with rule-based methods) given that most summarization datasets only have document-summary pairs. Since these labels might be suboptimal, we propose a latent variable extractive model where sentences are viewed as latent variables and sentences with activated variables are used to infer gold summaries. During training the loss comes \emph{directly} from gold summaries. Experiments on the CNN/Dailymail dataset show that our model improves over a strong extractive baseline trained on heuristically approximated labels and also performs competitively to several recent models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Xingxing Zhang (65 papers)
  2. Mirella Lapata (135 papers)
  3. Furu Wei (291 papers)
  4. Ming Zhou (182 papers)
Citations (152)