Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Summ^N: A Multi-Stage Summarization Framework for Long Input Dialogues and Documents (2110.10150v2)

Published 16 Oct 2021 in cs.CL

Abstract: Text summarization helps readers capture salient information from documents, news, interviews, and meetings. However, most state-of-the-art pretrained LLMs (LM) are unable to efficiently process long text for many summarization tasks. In this paper, we propose Summ$N$, a simple, flexible, and effective multi-stage framework for input texts that are longer than the maximum context length of typical pretrained LMs. Summ$N$ first splits the data samples and generates a coarse summary in multiple stages and then produces the final fine-grained summary based on it. Our framework can process input text of arbitrary length by adjusting the number of stages while keeping the LM input size fixed. Moreover, it can deal with both single-source documents and dialogues, and it can be used on top of different backbone abstractive summarization models. To the best of our knowledge, Summ$N$ is the first multi-stage split-then-summarize framework for long input summarization. Our experiments demonstrate that Summ$N$ outperforms previous state-of-the-art methods by improving ROUGE scores on three long meeting summarization datasets AMI, ICSI, and QMSum, two long TV series datasets from SummScreen, and a long document summarization dataset GovReport. Our data and code are available at https://github.com/psunlpgroup/Summ-N.

SummN^N: A Multi-Stage Summarization Framework for Long Input Dialogues and Documents

The paper "SummN^N: A Multi-Stage Summarization Framework for Long Input Dialogues and Documents" presents a novel approach to overcoming the challenges posed by the summarization of extended texts, which typically exceed the context length limits of current pretrained LMs. The authors introduce SummN^N, a multi-stage framework designed to split lengthy texts into manageable segments, allowing for the generation of comprehensive summaries without truncating context-relevant information.

Framework Overview

SummN^N leverages a multi-stage process, differentiating it as a pioneering method in the domain of long text summarization. The initial stages focus on dividing source texts into smaller, digestible segments before producing intermediary coarse summaries. This segmentation is crucial, as it maintains context dependency and allows all parts of the source text to contribute to the summary generation process. A greedy ROUGE-based algorithm aids in pairing these segments with appropriate target summaries, optimizing information retention.

In subsequent stages, SummN^N employs pre-trained abstractive summarization models to refine these coarse summaries into fine-grained versions. This approach effectively extends the receptive field of summarization models, allowing them to incorporate full context despite the original text's length. Notably, SummN^N can be adapted for both single-source documents and dialogues, showcasing its versatility across different text types.

Experimental Results

Experiments indicate that SummN^N yields superior ROUGE scores compared to existing methods across a diverse set of datasets, including AMI, ICSI, QMSum, SummScreen, and GovReport. The consistent improvement in summarization quality across these varied datasets underscores SummN^N's robustness and effectiveness. Additionally, SummN^N demonstrates significant enhancements over backbone models such as BART, T5, and PEGASUS, confirming the framework's capability to amplify pretrained models’ summarization performance on long input tasks.

Implications and Future Work

This paper provides significant insights into handling long document summarization, proposing mechanisms that efficiently utilize existing Transformer-based models. The ability to adapt various backbone models into the SummN^N framework suggests extensive applications across industries requiring detailed document synthesis, such as legal and technical fields.

Future research could explore optimizing the choice of coarse versus fine-grained stages based on dynamic context understanding, reinforcing the model’s adaptability to different text structures and types. Exploring inter-stage learning mechanisms and parameter sharing might yield efficiency improvements, particularly regarding computational resource utilization.

In conclusion, SummN^N represents a substantive advancement in the summarization of lengthy texts, with implications that extend both theoretical foundations and practical methodologies in AI-driven text processing.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Yusen Zhang (30 papers)
  2. Ansong Ni (17 papers)
  3. Ziming Mao (14 papers)
  4. Chen Henry Wu (17 papers)
  5. Chenguang Zhu (100 papers)
  6. Budhaditya Deb (11 papers)
  7. Ahmed H. Awadallah (7 papers)
  8. Dragomir Radev (98 papers)
  9. Rui Zhang (1138 papers)
Citations (83)