Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adaptive Contrastive Search: Uncertainty-Guided Decoding for Open-Ended Text Generation (2407.18698v2)

Published 26 Jul 2024 in cs.CL, cs.LG, stat.ME, and stat.ML

Abstract: Decoding from the output distributions of LLMs to produce high-quality text is a complex challenge in LLMing. Various approaches, such as beam search, sampling with temperature, $k-$sampling, nucleus $p-$sampling, typical decoding, contrastive decoding, and contrastive search, have been proposed to address this problem, aiming to improve coherence, diversity, as well as resemblance to human-generated text. In this study, we introduce adaptive contrastive search, a novel decoding strategy extending contrastive search by incorporating an adaptive degeneration penalty, guided by the estimated uncertainty of the model at each generation step. This strategy is designed to enhance both the creativity and diversity of the LLMing process while at the same time producing coherent and high-quality generated text output. Our findings indicate performance enhancement in both aspects, across different model architectures and datasets, underscoring the effectiveness of our method in text generation tasks. Our code base, datasets, and models are publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Esteban Garces Arias (9 papers)
  2. Julian Rodemann (17 papers)
  3. Meimingwei Li (4 papers)
  4. Christian Heumann (29 papers)
  5. Matthias Aßenmacher (20 papers)
Citations (2)
X Twitter Logo Streamline Icon: https://streamlinehq.com