ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2511.16854
176
0
v1v2v3 (latest)

MRI Super-Resolution with Deep Learning: A Comprehensive Survey

20 November 2025
Mohammad Khateri
Serge Vasylechko
Morteza Ghahremani
Liam Timms
Deniz Kocanaogullari
Simon K. Warfield
Camilo Jaimes
Davood Karimi
Alejandra Sierra
Jussi Tohka
Sila Kurugol
O. Afacan
ArXiv (abs)PDFHTMLGithub (10★)
Main:24 Pages
16 Figures
9 Tables
Appendix:17 Pages
Abstract

High-resolution (HR) magnetic resonance imaging (MRI) is crucial for many clinical and research applications. However, achieving it remains costly and constrained by technical trade-offs and experimental limitations. Super-resolution (SR) presents a promising computational approach to overcome these challenges by generating HR images from more affordable low-resolution (LR) scans, potentially improving diagnostic accuracy and efficiency without requiring additional hardware. This survey reviews recent advances in MRI SR techniques, with a focus on deep learning (DL) approaches. It examines DL-based MRI SR methods from the perspectives of computer vision, computational imaging, inverse problems, and MR physics, covering theoretical foundations, architectural designs, learning strategies, benchmark datasets, and performance metrics. We propose a systematic taxonomy to categorize these methods and present an in-depth study of both established and emerging SR techniques applicable to MRI, considering unique challenges in clinical and research contexts. We also highlight open challenges and directions that the community needs to address. Additionally, we provide a collection of essential open-access resources, tools, and tutorials, available on our GitHub: this https URL.

View on arXiv
Comments on this paper