Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Hierarchical Encoding-Decoding Scheme for Abstractive Multi-document Summarization (2305.08503v5)

Published 15 May 2023 in cs.CL

Abstract: Pre-trained LLMs (PLMs) have achieved outstanding achievements in abstractive single-document summarization (SDS). However, such benefits may not fully extend to multi-document summarization (MDS), where the handling of cross-document information is more complex. Previous works either design new MDS architectures or apply PLMs bluntly with concatenated source documents as a reformulated SDS task. While the former does not utilize previous pre-training efforts and may not generalize well across different domains, the latter may not sufficiently attend to the intricate cross-document relationships unique to MDS tasks. Instead, we enforce hierarchy on both the encoder and decoder to better utilize a PLM to facilitate multi-document interactions for the MDS task. Across 10 MDS benchmarks from various domains, our method outperforms or is competitive with the previous best models, including those with additional MDS pre-training or with more parameters. It outperforms its corresponding PLM backbone by up to 3 Rouge-L and is favored by humans.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Chenhui Shen (7 papers)
  2. Liying Cheng (16 papers)
  3. Xuan-Phi Nguyen (22 papers)
  4. Yang You (173 papers)
  5. Lidong Bing (144 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.