An Adversarial Example for Direct Logit Attribution: Memory Management in GELU-4L
An Adversarial Example for Direct Logit Attribution: Memory Management in GELU-4L
Prior work suggests that language models manage the limited bandwidth of the residual stream through a "memory management" mechanism, where certain attention heads and MLP layers clear residual stream directions set by earlier layers. Our study provides concrete evidence for this erasure phenomenon in a 4-layer transformer, identifying heads that consistently remove the output of earlier heads. We further demonstrate that direct logit attribution (DLA), a common technique for interpreting the output of intermediate transformer layers, can show misleading results by not accounting for erasure.
Can Rager、Jett Janiak、James Dao、Yeu-Tong Lau
10.18653/v1/2024.blackboxnlp-1.15
计算技术、计算机技术
Can Rager,Jett Janiak,James Dao,Yeu-Tong Lau.An Adversarial Example for Direct Logit Attribution: Memory Management in GELU-4L[EB/OL].(2023-10-11)[2025-05-18].https://arxiv.org/abs/2310.07325.点此复制
评论