Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Jointly Extracting and Compressing Documents with Summary State Representations (1904.02020v2)

Published 3 Apr 2019 in cs.IR, cs.CL, and cs.LG

Abstract: We present a new neural model for text summarization that first extracts sentences from a document and then compresses them. The proposed model offers a balance that sidesteps the difficulties in abstractive methods while generating more concise summaries than extractive methods. In addition, our model dynamically determines the length of the output summary based on the gold summaries it observes during training and does not require length constraints typical to extractive summarization. The model achieves state-of-the-art results on the CNN/DailyMail and Newsroom datasets, improving over current extractive and abstractive methods. Human evaluations demonstrate that our model generates concise and informative summaries. We also make available a new dataset of oracle compressive summaries derived automatically from the CNN/DailyMail reference summaries.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Afonso Mendes (10 papers)
  2. Shashi Narayan (35 papers)
  3. Sebastião Miranda (5 papers)
  4. Zita Marinho (15 papers)
  5. André F. T. Martins (113 papers)
  6. Shay B. Cohen (78 papers)
Citations (52)