ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.21715
59
0

Privacy-Preserving Chest X-ray Report Generation via Multimodal Federated Learning with ViT and GPT-2

27 May 2025
Md. Zahid Hossain
Mustofa Ahmed
Most. Sharmin Sultana Samu
Md. Rakibul Islam
    MedIm
ArXiv (abs)PDFHTML
Main:18 Pages
11 Figures
Bibliography:6 Pages
8 Tables
Abstract

The automated generation of radiology reports from chest X-ray images holds significant promise in enhancing diagnostic workflows while preserving patient privacy. Traditional centralized approaches often require sensitive data transfer, posing privacy concerns. To address this, the study proposes a Multimodal Federated Learning framework for chest X-ray report generation using the IU-Xray dataset. The system utilizes a Vision Transformer (ViT) as the encoder and GPT-2 as the report generator, enabling decentralized training without sharing raw data. Three Federated Learning (FL) aggregation strategies: FedAvg, Krum Aggregation and a novel Loss-aware Federated Averaging (L-FedAvg) were evaluated. Among these, Krum Aggregation demonstrated superior performance across lexical and semantic evaluation metrics such as ROUGE, BLEU, BERTScore and RaTEScore. The results show that FL can match or surpass centralized models in generating clinically relevant and semantically rich radiology reports. This lightweight and privacy-preserving framework paves the way for collaborative medical AI development without compromising data confidentiality.

View on arXiv
@article{hossain2025_2505.21715,
  title={ Privacy-Preserving Chest X-ray Report Generation via Multimodal Federated Learning with ViT and GPT-2 },
  author={ Md. Zahid Hossain and Mustofa Ahmed and Most. Sharmin Sultana Samu and Md. Rakibul Islam },
  journal={arXiv preprint arXiv:2505.21715},
  year={ 2025 }
}
Comments on this paper