LZ Penalty: An Information-Theoretic Repetition Penalty for Autoregressive Language Models.

Abstract

We introduce the Lempel-Ziv (LZ) penalty, a penalty specialized for reducing degenerate repetitions in autoregressive language models without loss of capability. The penalty is based on the codelengths in the LZ77 universal lossless compression algorithm. Through the lens of the prediction-compression duality, decoding with the LZ penalty has the interpretation of sampling from the residual distribution after removing the information that is highly compressible. We demonstrate the LZ penalty enables open-source reasoning models to operate with greedy decoding without loss of capability and without instances of degenerate repetition. In contrast, both the industry-standard frequency penalty and repetition penalty are ineffective, incurring degenerate repetition rates of up to 4% or more.

Cite

Text

Ginart et al. "LZ Penalty: An Information-Theoretic Repetition Penalty for Autoregressive Language Models.." Transactions on Machine Learning Research, 2026.

Markdown

[Ginart et al. "LZ Penalty: An Information-Theoretic Repetition Penalty for Autoregressive Language Models.." Transactions on Machine Learning Research, 2026.](https://mlanthology.org/tmlr/2026/ginart2026tmlr-lz/)

BibTeX

@article{ginart2026tmlr-lz,
  title     = {{LZ Penalty: An Information-Theoretic Repetition Penalty for Autoregressive Language Models.}},
  author    = {Ginart, Tony A and Kodali, Naveen and Lee, Jason and Xiong, Caiming and Savarese, Silvio and Emmons, John},
  journal   = {Transactions on Machine Learning Research},
  year      = {2026},
  url       = {https://mlanthology.org/tmlr/2026/ginart2026tmlr-lz/}
}