Papers
Topics
Authors
Recent
2000 character limit reached

LZ Penalty: An information-theoretic repetition penalty for autoregressive language models

Published 28 Apr 2025 in cs.LG, cs.AI, cs.IT, and math.IT | (2504.20131v2)

Abstract: We introduce the LZ penalty, a penalty specialized for reducing degenerate repetitions in autoregressive LLMs without loss of capability. The penalty is based on the codelengths in the LZ77 universal lossless compression algorithm. Through the lens of the prediction-compression duality, decoding the LZ penalty has the interpretation of sampling from the residual distribution after removing the information that is highly compressible. We demonstrate the LZ penalty enables state-of-the-art open-source reasoning models to operate with greedy (temperature zero) decoding without loss of capability and without instances of degenerate repetition. Both the industry-standard frequency penalty and repetition penalty are ineffective, incurring degenerate repetition rates of up to 4%.

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Paper to Video (Beta)

Open Problems

We found no open problems mentioned in this paper.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 7 tweets with 2529 likes about this paper.