|国家预印本平台
首页|GSsplat: Generalizable Semantic Gaussian Splatting for Novel-view Synthesis in 3D Scenes

GSsplat: Generalizable Semantic Gaussian Splatting for Novel-view Synthesis in 3D Scenes

GSsplat: Generalizable Semantic Gaussian Splatting for Novel-view Synthesis in 3D Scenes

来源:Arxiv_logoArxiv
英文摘要

The semantic synthesis of unseen scenes from multiple viewpoints is crucial for research in 3D scene understanding. Current methods are capable of rendering novel-view images and semantic maps by reconstructing generalizable Neural Radiance Fields. However, they often suffer from limitations in speed and segmentation performance. We propose a generalizable semantic Gaussian Splatting method (GSsplat) for efficient novel-view synthesis. Our model predicts the positions and attributes of scene-adaptive Gaussian distributions from once input, replacing the densification and pruning processes of traditional scene-specific Gaussian Splatting. In the multi-task framework, a hybrid network is designed to extract color and semantic information and predict Gaussian parameters. To augment the spatial perception of Gaussians for high-quality rendering, we put forward a novel offset learning module through group-based supervision and a point-level interaction module with spatial unit aggregation. When evaluated with varying numbers of multi-view inputs, GSsplat achieves state-of-the-art performance for semantic synthesis at the fastest speed.

Feng Xiao、Hongbin Xu、Wanlin Liang、Wenxiong Kang

计算技术、计算机技术

Feng Xiao,Hongbin Xu,Wanlin Liang,Wenxiong Kang.GSsplat: Generalizable Semantic Gaussian Splatting for Novel-view Synthesis in 3D Scenes[EB/OL].(2025-05-06)[2025-05-28].https://arxiv.org/abs/2505.04659.点此复制

评论