|国家预印本平台
首页|Physically-based Lighting Augmentation for Robotic Manipulation

Physically-based Lighting Augmentation for Robotic Manipulation

Physically-based Lighting Augmentation for Robotic Manipulation

来源:Arxiv_logoArxiv
英文摘要

Despite advances in data augmentation, policies trained via imitation learning still struggle to generalize across environmental variations such as lighting changes. To address this, we propose the first framework that leverages physically-based inverse rendering for lighting augmentation on real-world human demonstrations. Specifically, inverse rendering decomposes the first frame in each demonstration into geometric (surface normal, depth) and material (albedo, roughness, metallic) properties, which are then used to render appearance changes under different lighting. To ensure consistent augmentation across each demonstration, we fine-tune Stable Video Diffusion on robot execution videos for temporal lighting propagation. We evaluate our framework by measuring the structural and temporal consistency of the augmented sequences, and by assessing its effectiveness in reducing the behavior cloning generalization gap (40.1%) on a 7-DoF robot across 6 lighting conditions using 720 real-world evaluations. We further showcase three downstream applications enabled by the proposed framework.

Shutong Jin、Lezhong Wang、Ben Temming、Florian T. Pokorny

计算技术、计算机技术

Shutong Jin,Lezhong Wang,Ben Temming,Florian T. Pokorny.Physically-based Lighting Augmentation for Robotic Manipulation[EB/OL].(2025-08-02)[2025-08-19].https://arxiv.org/abs/2508.01442.点此复制

评论