24
0

Gaussian Mapping for Evolving Scenes

Main:9 Pages
3 Figures
Bibliography:3 Pages
10 Tables
Appendix:1 Pages
Abstract

Mapping systems with novel view synthesis (NVS) capabilities are widely used in computer vision, with augmented reality, robotics, and autonomous driving applications. Most notably, 3D Gaussian Splatting-based systems show high NVS performance; however, many current approaches are limited to static scenes. While recent works have started addressing short-term dynamics (motion within the view of the camera), long-term dynamics (the scene evolving through changes out of view) remain less explored. To overcome this limitation, we introduce a dynamic scene adaptation mechanism that continuously updates the 3D representation to reflect the latest changes. In addition, since maintaining geometric and semantic consistency remains challenging due to stale observations disrupting the reconstruction process, we propose a novel keyframe management mechanism that discards outdated observations while preserving as much information as possible. We evaluate Gaussian Mapping for Evolving Scenes (GaME) on both synthetic and real-world datasets and find it to be more accurate than the state of the art.

View on arXiv
@article{yugay2025_2506.06909,
  title={ Gaussian Mapping for Evolving Scenes },
  author={ Vladimir Yugay and Thies Kersten and Luca Carlone and Theo Gevers and Martin R. Oswald and Lukas Schmid },
  journal={arXiv preprint arXiv:2506.06909},
  year={ 2025 }
}
Comments on this paper