61
0

MRGAgents: A Multi-Agent Framework for Improved Medical Report Generation with Med-LVLMs

Main:7 Pages
3 Figures
3 Tables
Appendix:3 Pages
Abstract

Medical Large Vision-Language Models (Med-LVLMs) have been widely adopted for medical report generation. Despite Med-LVLMs producing state-of-the-art performance, they exhibit a bias toward predicting all findings as normal, leading to reports that overlook critical abnormalities. Furthermore, these models often fail to provide comprehensive descriptions of radiologically relevant regions necessary for accurate diagnosis. To address these challenges, we proposeMedical Report Generation Agents (MRGAgents), a novel multi-agent framework that fine-tunes specialized agents for different disease categories. By curating subsets of the IU X-ray and MIMIC-CXR datasets to train disease-specific agents, MRGAgents generates reports that more effectively balance normal and abnormal findings while ensuring a comprehensive description of clinically relevant regions. Our experiments demonstrate that MRGAgents outperformed the state-of-the-art, improving both report comprehensiveness and diagnostic utility.

View on arXiv
@article{wang2025_2505.18530,
  title={ MRGAgents: A Multi-Agent Framework for Improved Medical Report Generation with Med-LVLMs },
  author={ Pengyu Wang and Shuchang Ye and Usman Naseem and Jinman Kim },
  journal={arXiv preprint arXiv:2505.18530},
  year={ 2025 }
}
Comments on this paper