Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LaMSUM: Creating Extractive Summaries of User Generated Content using LLMs (2406.15809v2)

Published 22 Jun 2024 in cs.CL and cs.LG

Abstract: LLMs have demonstrated impressive performance across a wide range of NLP tasks, including summarization. LLMs inherently produce abstractive summaries by paraphrasing the original text, while the generation of extractive summaries - selecting specific subsets from the original text - remains largely unexplored. LLMs have a limited context window size, restricting the amount of data that can be processed at once. We tackle this challenge by introducing LaMSUM, a novel multi-level framework designed to generate extractive summaries from large collections of user-generated text using LLMs. LaMSUM integrates summarization with different voting methods to achieve robust summaries. Extensive evaluation using four popular LLMs (Llama 3, Mixtral, Gemini, GPT-4o) demonstrates that LaMSUM outperforms state-of-the-art extractive summarization methods. Overall, this work represents one of the first attempts to achieve extractive summarization by leveraging the power of LLMs, and is likely to spark further interest within the research community.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Garima Chhikara (3 papers)
  2. Anurag Sharma (6 papers)
  3. V. Gurucharan (4 papers)
  4. Kripabandhu Ghosh (34 papers)
  5. Abhijnan Chakraborty (35 papers)
Citations (1)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets