ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.14507
23
0

Can Pretrained Vision-Language Embeddings Alone Guide Robot Navigation?

17 June 2025
Nitesh Subedi
Adam Haroon
Shreyan Ganguly
Samuel T.K. Tetteh
Prajwal Koirala
Cody Fleming
Soumik Sarkar
    LM&Ro
ArXiv (abs)PDFHTML
Main:7 Pages
6 Figures
Bibliography:2 Pages
2 Tables
Abstract

Foundation models have revolutionized robotics by providing rich semantic representations without task-specific training. While many approaches integrate pretrained vision-language models (VLMs) with specialized navigation architectures, the fundamental question remains: can these pretrained embeddings alone successfully guide navigation without additional fine-tuning or specialized modules? We present a minimalist framework that decouples this question by training a behavior cloning policy directly on frozen vision-language embeddings from demonstrations collected by a privileged expert. Our approach achieves a 74% success rate in navigation to language-specified targets, compared to 100% for the state-aware expert, though requiring 3.2 times more steps on average. This performance gap reveals that pretrained embeddings effectively support basic language grounding but struggle with long-horizon planning and spatial reasoning. By providing this empirical baseline, we highlight both the capabilities and limitations of using foundation models as drop-in representations for embodied tasks, offering critical insights for robotics researchers facing practical design tradeoffs between system complexity and performance in resource-constrained scenarios. Our code is available atthis https URL

View on arXiv
@article{subedi2025_2506.14507,
  title={ Can Pretrained Vision-Language Embeddings Alone Guide Robot Navigation? },
  author={ Nitesh Subedi and Adam Haroon and Shreyan Ganguly and Samuel T.K. Tetteh and Prajwal Koirala and Cody Fleming and Soumik Sarkar },
  journal={arXiv preprint arXiv:2506.14507},
  year={ 2025 }
}
Comments on this paper