This paper studies the task of SatStreet-view synthesis, which aims to render photorealistic street-view panorama images and videos given any satellite image and specified camera positions or trajectories. We formulate to learn neural radiance field from paired images captured from satellite and street viewpoints, which comes to be a challenging learning problem due to the sparse-view natural and the extremely-large viewpoint changes between satellite and street-view images. We tackle the challenges based on a task-specific observation that street-view specific elements, including the sky and illumination effects are only visible in street-view panoramas, and present a novel approach Sat2Density++ to accomplish the goal of photo-realistic street-view panoramas rendering by modeling these street-view specific in neural networks. In the experiments, our method is testified on both urban and suburban scene datasets, demonstrating that Sat2Density++ is capable of rendering photorealistic street-view panoramas that are consistent across multiple views and faithful to the satellite image.
View on arXiv@article{qian2025_2505.17001, title={ Seeing through Satellite Images at Street Views }, author={ Ming Qian and Bin Tan and Qiuyu Wang and Xianwei Zheng and Hanjiang Xiong and Gui-Song Xia and Yujun Shen and Nan Xue }, journal={arXiv preprint arXiv:2505.17001}, year={ 2025 } }