ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.12362
19
71

MONAI Label: A framework for AI-assisted Interactive Labeling of 3D Medical Images

23 March 2022
A. Diaz-Pinto
Sachidanand Alle
V. Nath
Yucheng Tang
Alvin Ihsani
Muhammad Asad
Fernando Pérez-García
P. Mehta
Wenqi Li
Mona Flores
H. Roth
Tom Kamiel Magda Vercauteren
Daguang Xu
Prerna Dogra
Sebastien Ourselin
Andrew Feng
M. Jorge Cardoso
ArXivPDFHTML
Abstract

The lack of annotated datasets is a major bottleneck for training new task-specific supervised machine learning models, considering that manual annotation is extremely expensive and time-consuming. To address this problem, we present MONAI Label, a free and open-source framework that facilitates the development of applications based on artificial intelligence (AI) models that aim at reducing the time required to annotate radiology datasets. Through MONAI Label, researchers can develop AI annotation applications focusing on their domain of expertise. It allows researchers to readily deploy their apps as services, which can be made available to clinicians via their preferred user interface. Currently, MONAI Label readily supports locally installed (3D Slicer) and web-based (OHIF) frontends and offers two active learning strategies to facilitate and speed up the training of segmentation algorithms. MONAI Label allows researchers to make incremental improvements to their AI-based annotation application by making them available to other researchers and clinicians alike. Additionally, MONAI Label provides sample AI-based interactive and non-interactive labeling applications, that can be used directly off the shelf, as plug-and-play to any given dataset. Significant reduced annotation times using the interactive model can be observed on two public datasets.

View on arXiv
Comments on this paper