Neuron-Level Differentiation of Memorization and Generalization in Large Language Models
Neuron-Level Differentiation of Memorization and Generalization in Large Language Models
We investigate how Large Language Models (LLMs) distinguish between memorization and generalization at the neuron level. Through carefully designed tasks, we identify distinct neuron subsets responsible for each behavior. Experiments on both a GPT-2 model trained from scratch and a pretrained LLaMA-3.2 model fine-tuned with LoRA show consistent neuron-level specialization. We further demonstrate that inference-time interventions on these neurons can steer the model's behavior toward memorization or generalization. To assess robustness, we evaluate intra-task and inter-task consistency, confirming that these neuron-behavior associations reflect generalizable patterns rather than dataset-specific artifacts. Our findings reveal modular structure in LLMs and enable controlling memorization and generalization behaviors at inference time.
Ko-Wei Huang、Yi-Fu Fu、Ching-Yu Tsai、Yu-Chieh Tu、Tzu-Ling Cheng、Cheng-Yu Lin、Yi-Ting Yang、Heng-Yi Liu、Keng-Te Liao、Da-Cheng Juan、Shou-De Lin
计算技术、计算机技术
Ko-Wei Huang,Yi-Fu Fu,Ching-Yu Tsai,Yu-Chieh Tu,Tzu-Ling Cheng,Cheng-Yu Lin,Yi-Ting Yang,Heng-Yi Liu,Keng-Te Liao,Da-Cheng Juan,Shou-De Lin.Neuron-Level Differentiation of Memorization and Generalization in Large Language Models[EB/OL].(2025-07-09)[2025-08-02].https://arxiv.org/abs/2412.18497.点此复制
评论