|国家预印本平台
首页|Styleclone: Face Stylization with Diffusion Based Data Augmentation

Styleclone: Face Stylization with Diffusion Based Data Augmentation

Styleclone: Face Stylization with Diffusion Based Data Augmentation

来源:Arxiv_logoArxiv
英文摘要

We present StyleClone, a method for training image-to-image translation networks to stylize faces in a specific style, even with limited style images. Our approach leverages textual inversion and diffusion-based guided image generation to augment small style datasets. By systematically generating diverse style samples guided by both the original style images and real face images, we significantly enhance the diversity of the style dataset. Using this augmented dataset, we train fast image-to-image translation networks that outperform diffusion-based methods in speed and quality. Experiments on multiple styles demonstrate that our method improves stylization quality, better preserves source image content, and significantly accelerates inference. Additionally, we provide a systematic evaluation of the augmentation techniques and their impact on stylization performance.

Neeraj Matiyali、Siddharth Srivastava、Gaurav Sharma

计算技术、计算机技术

Neeraj Matiyali,Siddharth Srivastava,Gaurav Sharma.Styleclone: Face Stylization with Diffusion Based Data Augmentation[EB/OL].(2025-08-23)[2025-09-05].https://arxiv.org/abs/2508.17045.点此复制

评论