|国家预印本平台
首页|Concept-TRAK: Understanding how diffusion models learn concepts through concept-level attribution

Concept-TRAK: Understanding how diffusion models learn concepts through concept-level attribution

Concept-TRAK: Understanding how diffusion models learn concepts through concept-level attribution

来源:Arxiv_logoArxiv
英文摘要

While diffusion models excel at image generation, their growing adoption raises critical concerns around copyright issues and model transparency. Existing attribution methods identify training examples influencing an entire image, but fall short in isolating contributions to specific elements, such as styles or objects, that matter most to stakeholders. To bridge this gap, we introduce \emph{concept-level attribution} via a novel method called \emph{Concept-TRAK}. Concept-TRAK extends influence functions with two key innovations: (1) a reformulated diffusion training loss based on diffusion posterior sampling, enabling robust, sample-specific attribution; and (2) a concept-aware reward function that emphasizes semantic relevance. We evaluate Concept-TRAK on the AbC benchmark, showing substantial improvements over prior methods. Through diverse case studies--ranging from identifying IP-protected and unsafe content to analyzing prompt engineering and compositional learning--we demonstrate how concept-level attribution yields actionable insights for responsible generative AI development and governance.

Yonghyun Park、Chieh-Hsin Lai、Satoshi Hayakawa、Yuhta Takida、Naoki Murata、Wei-Hsiang Liao、Woosung Choi、Kin Wai Cheuk、Junghyun Koo、Yuki Mitsufuji

计算技术、计算机技术

Yonghyun Park,Chieh-Hsin Lai,Satoshi Hayakawa,Yuhta Takida,Naoki Murata,Wei-Hsiang Liao,Woosung Choi,Kin Wai Cheuk,Junghyun Koo,Yuki Mitsufuji.Concept-TRAK: Understanding how diffusion models learn concepts through concept-level attribution[EB/OL].(2025-07-09)[2025-07-18].https://arxiv.org/abs/2507.06547.点此复制

评论