109
1

RelationField: Relate Anything in Radiance Fields

Abstract

Neural radiance fields are an emerging 3D scene representation and recently even been extended to learn features for scene understanding by distilling open-vocabulary features from vision-language models. However, current method primarily focus on object-centric representations, supporting object segmentation or detection, while understanding semantic relationships between objects remains largely unexplored. To address this gap, we propose RelationField, the first method to extract inter-object relationships directly from neural radiance fields. RelationField represents relationships between objects as pairs of rays within a neural radiance field, effectively extending its formulation to include implicit relationship queries. To teach RelationField complex, open-vocabulary relationships, relationship knowledge is distilled from multi-modal LLMs. To evaluate RelationField, we solve open-vocabulary 3D scene graph generation tasks and relationship-guided instance segmentation, achieving state-of-the-art performance in both tasks. See the project website atthis https URL.

View on arXiv
@article{koch2025_2412.13652,
  title={ RelationField: Relate Anything in Radiance Fields },
  author={ Sebastian Koch and Johanna Wald and Mirco Colosi and Narunas Vaskevicius and Pedro Hermosilla and Federico Tombari and Timo Ropinski },
  journal={arXiv preprint arXiv:2412.13652},
  year={ 2025 }
}
Comments on this paper