Adaptive Visuo-Tactile Fusion with Predictive Force Attention for Dexterous Manipulation

Effectively utilizing multi-sensory data is important for robots to generalize across diverse tasks. However, the heterogeneous nature of these modalities makes fusion challenging. Existing methods propose strategies to obtain comprehensively fused features but often ignore the fact that each modality requires different levels of attention at different manipulation stages. To address this, we propose a force-guided attention fusion module that adaptively adjusts the weights of visual and tactile features without human labeling. We also introduce a self-supervised future force prediction auxiliary task to reinforce the tactile modality, improve data imbalance, and encourage proper adjustment. Our method achieves an average success rate of 93% across three fine-grained, contactrich tasks in real-world experiments. Further analysis shows that our policy appropriately adjusts attention to each modality at different manipulation stages. The videos can be viewed atthis https URL.
View on arXiv@article{li2025_2505.13982, title={ Adaptive Visuo-Tactile Fusion with Predictive Force Attention for Dexterous Manipulation }, author={ Jinzhou Li and Tianhao Wu and Jiyao Zhang and Zeyuan Chen and Haotian Jin and Mingdong Wu and Yujun Shen and Yaodong Yang and Hao Dong }, journal={arXiv preprint arXiv:2505.13982}, year={ 2025 } }