|国家预印本平台
首页|Out of Control -- Why Alignment Needs Formal Control Theory (and an Alignment Control Stack)

Out of Control -- Why Alignment Needs Formal Control Theory (and an Alignment Control Stack)

Out of Control -- Why Alignment Needs Formal Control Theory (and an Alignment Control Stack)

来源:Arxiv_logoArxiv
英文摘要

This position paper argues that formal optimal control theory should be central to AI alignment research, offering a distinct perspective from prevailing AI safety and security approaches. While recent work in AI safety and mechanistic interpretability has advanced formal methods for alignment, they often fall short of the generalisation required of control frameworks for other technologies. There is also a lack of research into how to render different alignment/control protocols interoperable. We argue that by recasting alignment through principles of formal optimal control and framing alignment in terms of hierarchical stack from physical to socio-technical layers according to which controls may be applied we can develop a better understanding of the potential and limitations for controlling frontier models and agentic AI systems. To this end, we introduce an Alignment Control Stack which sets out a hierarchical layered alignment stack, identifying measurement and control characteristics at each layer and how different layers are formally interoperable. We argue that such analysis is also key to the assurances that will be needed by governments and regulators in order to see AI technologies sustainably benefit the community. Our position is that doing so will bridge the well-established and empirically validated methods of optimal control with practical deployment considerations to create a more comprehensive alignment framework, enhancing how we approach safety and reliability for advanced AI systems.

Elija Perrier

自动化基础理论计算技术、计算机技术

Elija Perrier.Out of Control -- Why Alignment Needs Formal Control Theory (and an Alignment Control Stack)[EB/OL].(2025-06-21)[2025-07-21].https://arxiv.org/abs/2506.17846.点此复制

评论