107
0
v1v2 (latest)

Does Your 3D Encoder Really Work? When Pretrain-SFT from 2D VLMs Meets 3D VLMs

Main:9 Pages
10 Figures
Bibliography:3 Pages
17 Tables
Appendix:9 Pages
Abstract

Remarkable progress in 2D Vision-Language Models (VLMs) has spurred interest in extending them to 3D settings for tasks like 3D Question Answering, Dense Captioning, and Visual Grounding. Unlike 2D VLMs that typically process images through an image encoder, 3D scenes, with their intricate spatial structures, allow for diverse model architectures. Based on their encoder design, this paper categorizes recent 3D VLMs into 3D object-centric, 2D image-based, and 3D scene-centric approaches. Despite the architectural similarity of 3D scene-centric VLMs to their 2D counterparts, they have exhibited comparatively lower performance compared with the latest 3D object-centric and 2D image-based approaches. To understand this gap, we conduct an in-depth analysis, revealing that 3D scene-centric VLMs show limited reliance on the 3D scene encoder, and the pre-train stage appears less effective than in 2D VLMs. Furthermore, we observe that data scaling benefits are less pronounced on larger datasets. Our investigation suggests that while these models possess cross-modal alignment capabilities, they tend to over-rely on linguistic cues and overfit to frequent answer distributions, thereby diminishing the effective utilization of the 3D encoder. To address these limitations and encourage genuine 3D scene understanding, we introduce a novel 3D Relevance Discrimination QA dataset designed to disrupt shortcut learning and improve 3D understanding. Our findings highlight the need for advanced evaluation and improved strategies for better 3D understanding in 3D VLMs.

View on arXiv
@article{li2025_2506.05318,
  title={ Does Your 3D Encoder Really Work? When Pretrain-SFT from 2D VLMs Meets 3D VLMs },
  author={ Haoyuan Li and Yanpeng Zhou and Yufei Gao and Tao Tang and Jianhua Han and Yujie Yuan and Dave Zhenyu Chen and Jiawang Bian and Hang Xu and Xiaodan Liang },
  journal={arXiv preprint arXiv:2506.05318},
  year={ 2025 }
}
Comments on this paper