|国家预印本平台
首页|Position Paper: Rethinking Privacy in RL for Sequential Decision-making in the Age of LLMs

Position Paper: Rethinking Privacy in RL for Sequential Decision-making in the Age of LLMs

Position Paper: Rethinking Privacy in RL for Sequential Decision-making in the Age of LLMs

来源:Arxiv_logoArxiv
英文摘要

The rise of reinforcement learning (RL) in critical real-world applications demands a fundamental rethinking of privacy in AI systems. Traditional privacy frameworks, designed to protect isolated data points, fall short for sequential decision-making systems where sensitive information emerges from temporal patterns, behavioral strategies, and collaborative dynamics. Modern RL paradigms, such as federated RL (FedRL) and RL with human feedback (RLHF) in large language models (LLMs), exacerbate these challenges by introducing complex, interactive, and context-dependent learning environments that traditional methods do not address. In this position paper, we argue for a new privacy paradigm built on four core principles: multi-scale protection, behavioral pattern protection, collaborative privacy preservation, and context-aware adaptation. These principles expose inherent tensions between privacy, utility, and interpretability that must be navigated as RL systems become more pervasive in high-stakes domains like healthcare, autonomous vehicles, and decision support systems powered by LLMs. To tackle these challenges, we call for the development of new theoretical frameworks, practical mechanisms, and rigorous evaluation methodologies that collectively enable effective privacy protection in sequential decision-making systems.

Flint Xiaofeng Fan、Cheston Tan、Roger Wattenhofer、Yew-Soon Ong

自动化基础理论计算技术、计算机技术

Flint Xiaofeng Fan,Cheston Tan,Roger Wattenhofer,Yew-Soon Ong.Position Paper: Rethinking Privacy in RL for Sequential Decision-making in the Age of LLMs[EB/OL].(2025-04-15)[2025-05-24].https://arxiv.org/abs/2504.11511.点此复制

评论