Multimodal Federated Learning With Missing Modalities through Feature Imputation Network

Multimodal federated learning holds immense potential for collaboratively training models from multiple sources without sharing raw data, addressing both data scarcity and privacy concerns, two key challenges in healthcare. A major challenge in training multimodal federated models in healthcare is the presence of missing modalities due to multiple reasons, including variations in clinical practice, cost and accessibility constraints, retrospective data collection, privacy concerns, and occasional technical or human errors. Previous methods typically rely on publicly available real datasets or synthetic data to compensate for missing modalities. However, obtaining real datasets for every disease is impractical, and training generative models to synthesize missing modalities is computationally expensive and prone to errors due to the high dimensionality of medical data. In this paper, we propose a novel, lightweight, low-dimensional feature translator to reconstruct bottleneck features of the missing modalities. Our experiments on three different datasets (MIMIC-CXR, NIH Open-I, and CheXpert), in both homogeneous and heterogeneous settings consistently improve the performance of competitive baselines. The code and implementation details are available at:this https URL
View on arXiv@article{poudel2025_2505.20232, title={ Multimodal Federated Learning With Missing Modalities through Feature Imputation Network }, author={ Pranav Poudel and Aavash Chhetri and Prashnna Gyawali and Georgios Leontidis and Binod Bhattarai }, journal={arXiv preprint arXiv:2505.20232}, year={ 2025 } }