Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Simple Contrastive Learning Objective for Alleviating Neural Text Degeneration (2205.02517v2)

Published 5 May 2022 in cs.CL

Abstract: The cross-entropy objective has proved to be an all-purpose training objective for autoregressive LLMs (LMs). However, without considering the penalization of problematic tokens, LMs trained using cross-entropy exhibit text degeneration. To address this, unlikelihood training has been proposed to reduce the probability of unlikely tokens predicted by LMs. But unlikelihood does not consider the relationship between the label tokens and unlikely token candidates, thus showing marginal improvements in degeneration. We propose a new contrastive token learning objective that inherits the advantages of cross-entropy and unlikelihood training and avoids their limitations. The key idea is to teach a LM to generate high probabilities for label tokens and low probabilities of negative candidates. Comprehensive experiments on LLMing and open-domain dialogue generation tasks show that the proposed contrastive token objective yields much less repetitive texts, with a higher generation quality than baseline approaches, achieving the new state-of-the-art performance on text degeneration.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Shaojie Jiang (4 papers)
  2. Ruqing Zhang (60 papers)
  3. Svitlana Vakulenko (31 papers)
  4. Maarten de Rijke (263 papers)
Citations (15)