|国家预印本平台
首页|3D scene generation from scene graphs and self-attention

3D scene generation from scene graphs and self-attention

3D scene generation from scene graphs and self-attention

来源:Arxiv_logoArxiv
英文摘要

Synthesizing realistic and diverse indoor 3D scene layouts in a controllable fashion opens up applications in simulated navigation and virtual reality. As concise and robust representations of a scene, scene graphs have proven to be well-suited as the semantic control on the generated layout. We present a variant of the conditional variational autoencoder (cVAE) model to synthesize 3D scenes from scene graphs and floor plans. We exploit the properties of self-attention layers to capture high-level relationships between objects in a scene, and use these as the building blocks of our model. Our model, leverages graph transformers to estimate the size, dimension and orientation of the objects in a room while satisfying relationships in the given scene graph. Our experiments shows self-attention layers leads to sparser (7.9x compared to Graphto3D) and more diverse scenes (16%).

Diego Martin Arroyo、Federico Tombari、Davide Scaramuzza、Nico Messikomer、Mengqi Wang、Fabian Manhardt、Pietro Bonazzi

计算技术、计算机技术

Diego Martin Arroyo,Federico Tombari,Davide Scaramuzza,Nico Messikomer,Mengqi Wang,Fabian Manhardt,Pietro Bonazzi.3D scene generation from scene graphs and self-attention[EB/OL].(2024-04-02)[2025-08-02].https://arxiv.org/abs/2404.01887.点此复制

评论