|国家预印本平台
首页|Undesirable Memorization in Large Language Models: A Survey

Undesirable Memorization in Large Language Models: A Survey

Undesirable Memorization in Large Language Models: A Survey

来源:Arxiv_logoArxiv
英文摘要

While recent research increasingly showcases the remarkable capabilities of Large Language Models (LLMs), it is equally crucial to examine their associated risks. Among these, privacy and security vulnerabilities are particularly concerning, posing significant ethical and legal challenges. At the heart of these vulnerabilities stands memorization, which refers to a model's tendency to store and reproduce phrases from its training data. This phenomenon has been shown to be a fundamental source to various privacy and security attacks against LLMs. In this paper, we provide a taxonomy of the literature on LLM memorization, exploring it across three dimensions: granularity, retrievability, and desirability. Next, we discuss the metrics and methods used to quantify memorization, followed by an analysis of the causes and factors that contribute to memorization phenomenon. We then explore strategies that are used so far to mitigate the undesirable aspects of this phenomenon. We conclude our survey by identifying potential research topics for the near future, including methods to balance privacy and performance, and the analysis of memorization in specific LLM contexts such as conversational agents, retrieval-augmented generation, and diffusion language models. Given the rapid research pace in this field, we also maintain a dedicated repository of the references discussed in this survey which will be regularly updated to reflect the latest developments.

Suzan Verberne、Ali Satvaty、Fatih Turkmen

计算技术、计算机技术

Suzan Verberne,Ali Satvaty,Fatih Turkmen.Undesirable Memorization in Large Language Models: A Survey[EB/OL].(2024-10-03)[2025-05-07].https://arxiv.org/abs/2410.02650.点此复制

评论