|国家预印本平台
首页|A Controllable Appearance Representation for Flexible Transfer and Editing

A Controllable Appearance Representation for Flexible Transfer and Editing

A Controllable Appearance Representation for Flexible Transfer and Editing

来源:Arxiv_logoArxiv
英文摘要

We present a method that computes an interpretable representation of material appearance within a highly compact, disentangled latent space. This representation is learned in a self-supervised fashion using an adapted FactorVAE. We train our model with a carefully designed unlabeled dataset, avoiding possible biases induced by human-generated labels. Our model demonstrates strong disentanglement and interpretability by effectively encoding material appearance and illumination, despite the absence of explicit supervision. Then, we use our representation as guidance for training a lightweight IP-Adapter to condition a diffusion pipeline that transfers the appearance of one or more images onto a target geometry, and allows the user to further edit the resulting appearance. Our approach offers fine-grained control over the generated results: thanks to the well-structured compact latent space, users can intuitively manipulate attributes such as hue or glossiness in image space to achieve the desired final appearance.

Santiago Jimenez-Navarro、Julia Guerrero-Viu、Belen Masia

计算技术、计算机技术

Santiago Jimenez-Navarro,Julia Guerrero-Viu,Belen Masia.A Controllable Appearance Representation for Flexible Transfer and Editing[EB/OL].(2025-04-21)[2025-05-09].https://arxiv.org/abs/2504.15028.点此复制

评论