Ownership Verification of DNN Models Using White-Box Adversarial Attacks with Specified Probability Manipulation
Ownership Verification of DNN Models Using White-Box Adversarial Attacks with Specified Probability Manipulation
In this paper, we propose a novel framework for ownership verification of deep neural network (DNN) models for image classification tasks. It allows verification of model identity by both the rightful owner and third party without presenting the original model. We assume a gray-box scenario where an unauthorized user owns a model that is illegally copied from the original model, provides services in a cloud environment, and the user throws images and receives the classification results as a probability distribution of output classes. The framework applies a white-box adversarial attack to align the output probability of a specific class to a designated value. Due to the knowledge of original model, it enables the owner to generate such adversarial examples. We propose a simple but effective adversarial attack method based on the iterative Fast Gradient Sign Method (FGSM) by introducing control parameters. Experimental results confirm the effectiveness of the identification of DNN models using adversarial attack.
Teruki Sano、Minoru Kuribayashi、Masao Sakai、Shuji Ishobe、Eisuke Koizumi
计算技术、计算机技术
Teruki Sano,Minoru Kuribayashi,Masao Sakai,Shuji Ishobe,Eisuke Koizumi.Ownership Verification of DNN Models Using White-Box Adversarial Attacks with Specified Probability Manipulation[EB/OL].(2025-05-23)[2025-06-07].https://arxiv.org/abs/2505.17579.点此复制
评论