ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.03792
65
0

Rebalanced Multimodal Learning with Data-aware Unimodal Sampling

5 March 2025
Qingyuan Jiang
Zhouyang Chi
Xiao Ma
Qirong Mao
Yang Yang
Jinhui Tang
ArXivPDFHTML
Abstract

To address the modality learning degeneration caused by modality imbalance, existing multimodal learning~(MML) approaches primarily attempt to balance the optimization process of each modality from the perspective of model learning. However, almost all existing methods ignore the modality imbalance caused by unimodal data sampling, i.e., equal unimodal data sampling often results in discrepancies in informational content, leading to modality imbalance. Therefore, in this paper, we propose a novel MML approach called \underline{D}ata-aware \underline{U}nimodal \underline{S}ampling~(\method), which aims to dynamically alleviate the modality imbalance caused by sampling. Specifically, we first propose a novel cumulative modality discrepancy to monitor the multimodal learning process. Based on the learning status, we propose a heuristic and a reinforcement learning~(RL)-based data-aware unimodal sampling approaches to adaptively determine the quantity of sampled data at each iteration, thus alleviating the modality imbalance from the perspective of sampling. Meanwhile, our method can be seamlessly incorporated into almost all existing multimodal learning approaches as a plugin. Experiments demonstrate that \method~can achieve the best performance by comparing with diverse state-of-the-art~(SOTA) baselines.

View on arXiv
@article{jiang2025_2503.03792,
  title={ Rebalanced Multimodal Learning with Data-aware Unimodal Sampling },
  author={ Qingyuan Jiang and Zhouyang Chi and Xiao Ma and Qirong Mao and Yang Yang and Jinhui Tang },
  journal={arXiv preprint arXiv:2503.03792},
  year={ 2025 }
}
Comments on this paper