ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.07455
29
0

GelFusion: Enhancing Robotic Manipulation under Visual Constraints via Visuotactile Fusion

12 May 2025
Shulong Jiang
Shiqi Zhao
Yuxuan Fan
Peng Yin
ArXivPDFHTML
Abstract

Visuotactile sensing offers rich contact information that can help mitigate performance bottlenecks in imitation learning, particularly under vision-limited conditions, such as ambiguous visual cues or occlusions. Effectively fusing visual and visuotactile modalities, however, presents ongoing challenges. We introduce GelFusion, a framework designed to enhance policies by integrating visuotactile feedback, specifically from high-resolution GelSight sensors. GelFusion using a vision-dominated cross-attention fusion mechanism incorporates visuotactile information into policy learning. To better provide rich contact information, the framework's core component is our dual-channel visuotactile feature representation, simultaneously leveraging both texture-geometric and dynamic interaction features. We evaluated GelFusion on three contact-rich tasks: surface wiping, peg insertion, and fragile object pick-and-place. Outperforming baselines, GelFusion shows the value of its structure in improving the success rate of policy learning.

View on arXiv
@article{jiang2025_2505.07455,
  title={ GelFusion: Enhancing Robotic Manipulation under Visual Constraints via Visuotactile Fusion },
  author={ Shulong Jiang and Shiqi Zhao and Yuxuan Fan and Peng Yin },
  journal={arXiv preprint arXiv:2505.07455},
  year={ 2025 }
}
Comments on this paper