Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GRETEL: Graph Contrastive Topic Enhanced Language Model for Long Document Extractive Summarization (2208.09982v1)

Published 21 Aug 2022 in cs.CL

Abstract: Recently, neural topic models (NTMs) have been incorporated into pre-trained LLMs (PLMs), to capture the global semantic information for text summarization. However, in these methods, there remain limitations in the way they capture and integrate the global semantic information. In this paper, we propose a novel model, the graph contrastive topic enhanced LLM (GRETEL), that incorporates the graph contrastive topic model with the pre-trained LLM, to fully leverage both the global and local contextual semantics for long document extractive summarization. To better capture and incorporate the global semantic information into PLMs, the graph contrastive topic model integrates the hierarchical transformer encoder and the graph contrastive learning to fuse the semantic information from the global document context and the gold summary. To this end, GRETEL encourages the model to efficiently extract salient sentences that are topically related to the gold summary, rather than redundant sentences that cover sub-optimal topics. Experimental results on both general domain and biomedical datasets demonstrate that our proposed method outperforms SOTA methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Qianqian Xie (60 papers)
  2. Jimin Huang (37 papers)
  3. Tulika Saha (3 papers)
  4. Sophia Ananiadou (72 papers)
Citations (23)

Summary

We haven't generated a summary for this paper yet.