|国家预印本平台
首页|Coherent 3D Scene Diffusion From a Single RGB Image

Coherent 3D Scene Diffusion From a Single RGB Image

Coherent 3D Scene Diffusion From a Single RGB Image

来源:Arxiv_logoArxiv
英文摘要

We present a novel diffusion-based approach for coherent 3D scene reconstruction from a single RGB image. Our method utilizes an image-conditioned 3D scene diffusion model to simultaneously denoise the 3D poses and geometries of all objects within the scene. Motivated by the ill-posed nature of the task and to obtain consistent scene reconstruction results, we learn a generative scene prior by conditioning on all scene objects simultaneously to capture the scene context and by allowing the model to learn inter-object relationships throughout the diffusion process. We further propose an efficient surface alignment loss to facilitate training even in the absence of full ground-truth annotation, which is common in publicly available datasets. This loss leverages an expressive shape representation, which enables direct point sampling from intermediate shape predictions. By framing the task of single RGB image 3D scene reconstruction as a conditional diffusion process, our approach surpasses current state-of-the-art methods, achieving a 12.04% improvement in AP3D on SUN RGB-D and a 13.43% increase in F-Score on Pix3D.

Manuel Dahnert、Matthias Nie?ner、Norman M¨1ller、Angela Dai

计算技术、计算机技术

Manuel Dahnert,Matthias Nie?ner,Norman M¨1ller,Angela Dai.Coherent 3D Scene Diffusion From a Single RGB Image[EB/OL].(2024-12-13)[2025-08-02].https://arxiv.org/abs/2412.10294.点此复制

评论