WATCH: Adaptive Monitoring for AI Deployments via Weighted-Conformal Martingales
WATCH: Adaptive Monitoring for AI Deployments via Weighted-Conformal Martingales
Responsibly deploying artificial intelligence (AI) / machine learning (ML) systems in high-stakes settings arguably requires not only proof of system reliability, but moreover continual, post-deployment monitoring to quickly detect and address any unsafe behavior. Statistical methods for nonparametric change-point detection -- especially the tools of conformal test martingales (CTMs) and anytime-valid inference -- offer promising approaches to this monitoring task. However, existing methods are restricted to monitoring limited hypothesis classes or ``alarm criteria'' (such as data shifts that violate certain exchangeability assumptions), do not allow for online adaptation in response to shifts, and/or do not enable root-cause analysis of any degradation. In this paper, we expand the scope of these monitoring methods by proposing a weighted generalization of conformal test martingales (WCTMs), which lay a theoretical foundation for online monitoring for any unexpected changepoints in the data distribution while controlling false-alarms. For practical applications, we propose specific WCTM algorithms that adapt online to mild covariate shifts (in the marginal input distribution) while quickly detecting and diagnosing more severe shifts, such as concept shifts (in the conditional label distribution) or extreme (out-of-support) covariate shifts that cannot be easily adapted to. On real-world datasets, we demonstrate improved performance relative to state-of-the-art baselines.
Drew Prinster、Xing Han、Anqi Liu、Suchi Saria
计算技术、计算机技术
Drew Prinster,Xing Han,Anqi Liu,Suchi Saria.WATCH: Adaptive Monitoring for AI Deployments via Weighted-Conformal Martingales[EB/OL].(2025-05-07)[2025-05-18].https://arxiv.org/abs/2505.04608.点此复制
评论