Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

COLO: A Contrastive Learning based Re-ranking Framework for One-Stage Summarization (2209.14569v2)

Published 29 Sep 2022 in cs.CL

Abstract: Traditional training paradigms for extractive and abstractive summarization systems always only use token-level or sentence-level training objectives. However, the output summary is always evaluated from summary-level which leads to the inconsistency in training and evaluation. In this paper, we propose a Contrastive Learning based re-ranking framework for one-stage summarization called COLO. By modeling a contrastive objective, we show that the summarization model is able to directly generate summaries according to the summary-level score without additional modules and parameters. Extensive experiments demonstrate that COLO boosts the extractive and abstractive results of one-stage systems on CNN/DailyMail benchmark to 44.58 and 46.33 ROUGE-1 score while preserving the parameter efficiency and inference efficiency. Compared with state-of-the-art multi-stage systems, we save more than 100 GPU training hours and obtaining 3~8 speed-up ratio during inference while maintaining comparable results.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Chenxin An (17 papers)
  2. Ming Zhong (88 papers)
  3. Zhiyong Wu (171 papers)
  4. Qin Zhu (11 papers)
  5. Xuanjing Huang (287 papers)
  6. Xipeng Qiu (257 papers)
Citations (19)

Summary

We haven't generated a summary for this paper yet.