|国家预印本平台
首页|LZ Penalty: An information-theoretic repetition penalty for autoregressive language models

LZ Penalty: An information-theoretic repetition penalty for autoregressive language models

LZ Penalty: An information-theoretic repetition penalty for autoregressive language models

来源:Arxiv_logoArxiv
英文摘要

We introduce the LZ penalty, a penalty specialized for reducing degenerate repetitions in autoregressive language models without loss of capability. The penalty is based on the codelengths in the LZ77 universal lossless compression algorithm. Through the lens of the prediction-compression duality, decoding the LZ penalty has the interpretation of sampling from the residual distribution after removing the information that is highly compressible. We demonstrate the LZ penalty enables state-of-the-art open-source reasoning models to operate with greedy (temperature zero) decoding without loss of capability and without instances of degenerate repetition. Both the industry-standard frequency penalty and repetition penalty are ineffective, incurring degenerate repetition rates of up to 4%.

Antonio A. Ginart、Naveen Kodali、Jason Lee、Caiming Xiong、Silvio Savarese、John R. Emmons

计算技术、计算机技术

Antonio A. Ginart,Naveen Kodali,Jason Lee,Caiming Xiong,Silvio Savarese,John R. Emmons.LZ Penalty: An information-theoretic repetition penalty for autoregressive language models[EB/OL].(2025-04-28)[2025-06-13].https://arxiv.org/abs/2504.20131.点此复制

评论