Few-Shot Class-Incremental Model Attribution Using Learnable Representation From CLIP-ViT Features
Few-Shot Class-Incremental Model Attribution Using Learnable Representation From CLIP-ViT Features
Recently, images that distort or fabricate facts using generative models have become a social concern. To cope with continuous evolution of generative artificial intelligence (AI) models, model attribution (MA) is necessary beyond just detection of synthetic images. However, current deep learning-based MA methods must be trained from scratch with new data to recognize unseen models, which is time-consuming and data-intensive. This work proposes a new strategy to deal with persistently emerging generative models. We adapt few-shot class-incremental learning (FSCIL) mechanisms for MA problem to uncover novel generative AI models. Unlike existing FSCIL approaches that focus on object classification using high-level information, MA requires analyzing low-level details like color and texture in synthetic images. Thus, we utilize a learnable representation from different levels of CLIP-ViT features. To learn an effective representation, we propose Adaptive Integration Module (AIM) to calculate a weighted sum of CLIP-ViT block features for each image, enhancing the ability to identify generative models. Extensive experiments show our method effectively extends from prior generative models to recent ones.
Hanbyul Lee、Juneho Yi
计算技术、计算机技术
Hanbyul Lee,Juneho Yi.Few-Shot Class-Incremental Model Attribution Using Learnable Representation From CLIP-ViT Features[EB/OL].(2025-03-11)[2025-04-26].https://arxiv.org/abs/2503.08148.点此复制
评论