ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.12379
75
0

OCT Data is All You Need: How Vision Transformers with and without Pre-training Benefit Imaging

17 February 2025
Zihao Han
Philippe De Wilde
    ViT
ArXivPDFHTML
Abstract

Optical Coherence Tomography (OCT) provides high-resolution cross-sectional images useful for diagnosing various diseases, but their distinct characteristics from natural images raise questions about whether large-scale pre-training on datasets like ImageNet is always beneficial. In this paper, we investigate the impact of ImageNet-based pre-training on Vision Transformer (ViT) performance for OCT image classification across different dataset sizes. Our experiments cover four-category retinal pathologies (CNV, DME, Drusen, Normal). Results suggest that while pre-training can accelerate convergence and potentially offer better performance in smaller datasets, training from scratch may achieve comparable or even superior accuracy when sufficient OCT data is available. Our findings highlight the importance of matching domain characteristics in pre-training and call for further study on large-scale OCT-specific pre-training.

View on arXiv
@article{han2025_2502.12379,
  title={ OCT Data is All You Need: How Vision Transformers with and without Pre-training Benefit Imaging },
  author={ Zihao Han and Philippe De Wilde },
  journal={arXiv preprint arXiv:2502.12379},
  year={ 2025 }
}
Comments on this paper