|国家预印本平台
首页|'Neural howlround' in large language models: a self-reinforcing bias phenomenon, and a dynamic attenuation solution

'Neural howlround' in large language models: a self-reinforcing bias phenomenon, and a dynamic attenuation solution

'Neural howlround' in large language models: a self-reinforcing bias phenomenon, and a dynamic attenuation solution

来源:Arxiv_logoArxiv
英文摘要

Large language model (LLM)-driven AI systems may exhibit an inference failure mode we term `neural howlround,' a self-reinforcing cognitive loop where certain highly weighted inputs become dominant, leading to entrenched response patterns resistant to correction. This paper explores the mechanisms underlying this phenomenon, which is distinct from model collapse and biased salience weighting. We propose an attenuation-based correction mechanism that dynamically introduces counterbalancing adjustments and can restore adaptive reasoning, even in `locked-in' AI systems. Additionally, we discuss some other related effects arising from improperly managed reinforcement. Finally, we outline potential applications of this mitigation strategy for improving AI robustness in real-world decision-making tasks.

Seth Drake

计算技术、计算机技术

Seth Drake.'Neural howlround' in large language models: a self-reinforcing bias phenomenon, and a dynamic attenuation solution[EB/OL].(2025-04-07)[2025-05-29].https://arxiv.org/abs/2504.07992.点此复制

评论