|国家预印本平台
首页|Personalized Federated Training of Diffusion Models with Privacy Guarantees

Personalized Federated Training of Diffusion Models with Privacy Guarantees

Personalized Federated Training of Diffusion Models with Privacy Guarantees

来源:Arxiv_logoArxiv
英文摘要

The scarcity of accessible, compliant, and ethically sourced data presents a considerable challenge to the adoption of artificial intelligence (AI) in sensitive fields like healthcare, finance, and biomedical research. Furthermore, access to unrestricted public datasets is increasingly constrained due to rising concerns over privacy, copyright, and competition. Synthetic data has emerged as a promising alternative, and diffusion models -- a cutting-edge generative AI technology -- provide an effective solution for generating high-quality and diverse synthetic data. In this paper, we introduce a novel federated learning framework for training diffusion models on decentralized private datasets. Our framework leverages personalization and the inherent noise in the forward diffusion process to produce high-quality samples while ensuring robust differential privacy guarantees. Our experiments show that our framework outperforms non-collaborative training methods, particularly in settings with high data heterogeneity, and effectively reduces biases and imbalances in synthetic data, resulting in fairer downstream models.

Kumar Kshitij Patel、Weitong Zhang、Lingxiao Wang

医学研究方法生物科学研究方法、生物科学研究技术

Kumar Kshitij Patel,Weitong Zhang,Lingxiao Wang.Personalized Federated Training of Diffusion Models with Privacy Guarantees[EB/OL].(2025-04-01)[2025-06-23].https://arxiv.org/abs/2504.00952.点此复制

评论