ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.02824
36
0

Developing a PET/CT Foundation Model for Cross-Modal Anatomical and Functional Imaging

4 March 2025
Y. Oh
Robert Seifert
Yihan Cao
Christoph Clement
Justin Ferdinandus
Constantin Lapa
Alessandro Liebich
Michelle Amon
Johanna Enke
S. Song
Runqi Meng
Fang Zeng
Ning Guo
X. Li
P. Heidari
Axel Rominger
Kuangyu Shi
Quanzheng Li
    ViT
    MedIm
ArXivPDFHTML
Abstract

In oncology, Positron Emission Tomography-Computed Tomography (PET/CT) is widely used in cancer diagnosis, staging, and treatment monitoring, as it combines anatomical details from CT with functional metabolic activity and molecular marker expression information from PET. However, existing artificial intelligence-driven PET/CT analyses rely predominantly on task-specific models trained from scratch or on limited datasets, limiting their generalizability and robustness. To address this, we propose a foundation model approach specifically designed for multimodal PET/CT imaging. We introduce the Cross-Fraternal Twin Masked Autoencoder (FratMAE), a novel framework that effectively integrates whole-body anatomical and functional or molecular information. FratMAE employs separate Vision Transformer (ViT) encoders for PET and CT scans, along with cross-attention decoders that enable synergistic interactions between modalities during masked autoencoder training. Additionally, it incorporates textual metadata to enhance PET representation learning. By pre-training on PET/CT datasets, FratMAE captures intricate cross-modal relationships and global uptake patterns, achieving superior performance on downstream tasks and demonstrating its potential as a generalizable foundation model.

View on arXiv
@article{oh2025_2503.02824,
  title={ Developing a PET/CT Foundation Model for Cross-Modal Anatomical and Functional Imaging },
  author={ Yujin Oh and Robert Seifert and Yihan Cao and Christoph Clement and Justin Ferdinandus and Constantin Lapa and Alessandro Liebich and Michelle Amon and Johanna Enke and Sifan Song and Runqi Meng and Fang Zeng and Ning Guo and Xiang Li and Pedram Heidari and Axel Rominger and Kuangyu Shi and Quanzheng Li },
  journal={arXiv preprint arXiv:2503.02824},
  year={ 2025 }
}
Comments on this paper