Multi-organ segmentation: a progressive exploration of learning paradigms under scarce annotation.

Phys Med Biol

Digital Medical Research Center, School of Basic Medical Science, Fudan University, Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, People's Republic of China.

Published: May 2024

Precise delineation of multiple organs or abnormal regions in the human body from medical images plays an essential role in computer-aided diagnosis, surgical simulation, image-guided interventions, and especially in radiotherapy treatment planning. Thus, it is of great significance to explore automatic segmentation approaches, among which deep learning-based approaches have evolved rapidly and witnessed remarkable progress in multi-organ segmentation. However, obtaining an appropriately sized and fine-grained annotated dataset of multiple organs is extremely hard and expensive. Such scarce annotation limits the development of high-performance multi-organ segmentation models but promotes many annotation-efficient learning paradigms. Among these, studies on transfer learning leveraging external datasets, semi-supervised learning including unannotated datasets and partially-supervised learning integrating partially-labeled datasets have led the dominant way to break such dilemmas in multi-organ segmentation. We first review the fully supervised method, then present a comprehensive and systematic elaboration of the 3 abovementioned learning paradigms in the context of multi-organ segmentation from both technical and methodological perspectives, and finally summarize their challenges and future trends.

Download full-text PDF

Source
http://dx.doi.org/10.1088/1361-6560/ad33b5DOI Listing

Publication Analysis

Top Keywords

multi-organ segmentation
20
learning paradigms
12
scarce annotation
8
multiple organs
8
learning
6
multi-organ
5
segmentation
5
segmentation progressive
4
progressive exploration
4
exploration learning
4

Similar Publications

Objective: This study aims to assess and compare two state-of-the-art deep learning approaches for segmenting four thoracic organs at risk (OAR)-the esophagus, trachea, heart, and aorta-in CT images in the context of radiotherapy planning.

Materials And Methods: We compare a multi-organ segmentation approach and the fusion of multiple single-organ models, each dedicated to one OAR. All were trained using nnU-Net with the default parameters and the full-resolution configuration.

View Article and Find Full Text PDF

Adaptive cascade decoders for segmenting challenging regions in medical images.

Comput Biol Med

December 2024

Aerospace Hi-tech Holding Group Co., LTD, Harbin, Heilongjiang, 150060, China.

CNN-based techniques have achieved impressive outcomes in medical image segmentation but struggle to capture long-term dependencies between pixels. The Transformer, with its strong feature extraction and representation learning abilities, performs exceptionally well within the domain of medical image partitioning. However, there are still shortcomings in bridging local to global connections, resulting in occasional loss of positional information.

View Article and Find Full Text PDF

Convolutional neural networks (CNNs) are well established in handling local features in visual tasks; yet, they falter in managing complex spatial relationships and long-range dependencies that are crucial for medical image segmentation, particularly in identifying pathological changes. While vision transformer (ViT) excels in addressing long-range dependencies, their ability to leverage local features remains inadequate. Recent ViT variants have merged CNNs to improve feature representation and segmentation outcomes, yet challenges with limited receptive fields and precise feature representation persist.

View Article and Find Full Text PDF

Background: Medical image segmentation is crucial for clinical diagnostics and treatment planning. Recently, hybrid models often neglect the local modeling capabilities of Transformers for medical image segmentation, despite the complementary nature of local information from both convolutional neural networks (CNNs) and transformers. This limitation is particularly problematic in multi-organ segmentation, where organs are closely adhered, and accurate delineation is essential.

View Article and Find Full Text PDF

TAC-UNet: transformer-assisted convolutional neural network for medical image segmentation.

Quant Imaging Med Surg

December 2024

The College of Computer and Information Science, Southwest University, Chongqing, China.

Background: Medical image segmentation is crucial for improving healthcare outcomes. Convolutional neural networks (CNNs) have been widely applied in medical image analysis; however, their inherent inductive biases limit their ability to capture global contextual information. Vision transformer (ViT) architectures address this limitation by leveraging attention mechanisms to model global relationships; however, they typically require large-scale datasets for effective training, which is challenging in the field of medical imaging due to limited data availability.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!