Privacy-Preserving Model and Preprocessing Verification for Machine Learning
Privacy-Preserving Model and Preprocessing Verification for Machine Learning
This paper presents a framework for privacy-preserving verification of machine learning models, focusing on models trained on sensitive data. Integrating Local Differential Privacy (LDP) with model explanations from LIME and SHAP, our framework enables robust verification without compromising individual privacy. It addresses two key tasks: binary classification, to verify if a target model was trained correctly by applying the appropriate preprocessing steps, and multi-class classification, to identify specific preprocessing errors. Evaluations on three real-world datasets-Diabetes, Adult, and Student Record-demonstrate that while the ML-based approach is particularly effective in binary tasks, the threshold-based method performs comparably in multi-class tasks. Results indicate that although verification accuracy varies across datasets and noise levels, the framework provides effective detection of preprocessing errors, strong privacy guarantees, and practical applicability for safeguarding sensitive data.
Anisa Halimi、Erman Ayday、Wenbiao Li、Jaideep Vaidya、Xiaoqian Jiang
计算技术、计算机技术
Anisa Halimi,Erman Ayday,Wenbiao Li,Jaideep Vaidya,Xiaoqian Jiang.Privacy-Preserving Model and Preprocessing Verification for Machine Learning[EB/OL].(2025-01-14)[2025-08-03].https://arxiv.org/abs/2501.08236.点此复制
评论