This paper presents a fully automatic method for multi-organ segmentation from 3D abdominal CT volumes. Firstly, spines and ribs are removed by exponential transformation and binarization to reduce the disturbance to subsequent segmentation. Then, a Local Linear Embedding (LLE)-based graph partitioning approach is employed to perform initial segmentation for liver, spleen, and bilateral kidneys simultaneously, and a novel segmentation refinement scheme is applied composed of hybrid intensity model, 3D Chan-Vese model, and histogram equalization-based organ separation algorithm. Finally, a pseudo-3D bottleneck detection algorithm is introduced for boundary correction. The proposed method does not require heavy training or registration process and is capable of dealing with shape and location variations as well as the weak boundaries of target organs. Experiments on XHCSU20 database show the proposed method is competitive with state-of-the-art methods with Dice similarity coefficients of 95.9%, 95.1%, 94.7%, and 94.5%, Jaccard indices of 92.2%, 90.8%, 90.0%, and 89.5%, and average symmetric surface distances of 1.1 mm, 1.0 mm, 0.9 mm and 1.1 mm, for liver, spleen, left and right kidneys, respectively, and the average running time is around 6 min for a CT volume. The accuracy, precision, recall, and specificity also maintain high values for each of the four organs. Moreover, experiments on SLIVER07 dataset prove its high efficiency and accuracy on liver-only segmentation.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1016/j.compbiomed.2021.105030 | DOI Listing |
Sci Data
January 2025
Department of Ophthalmology, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
This study presents TOM500, a comprehensive multi-organ annotated orbital magnetic resonance imaging (MRI) dataset. It includes clinical data, T2-weighted MRI scans, and corresponding segmentations from 500 patients with thyroid eye disease (TED) during their initial visit. TED is a common autoimmune disorder with distinct orbital MRI features.
View Article and Find Full Text PDFPhys Imaging Radiat Oncol
October 2024
Université Paris-Saclay, Gustave Roussy, Inserm, Molecular Radiotherapy and Therapeutic Innovation, U1030, 94800 Villejuif, France.
Background And Purpose: Deep-learning-based automatic segmentation is widely used in radiation oncology to delineate organs-at-risk. Dual-energy CT (DECT) allows the reconstruction of enhanced contrast images that could help with manual and auto-delineation. This paper presents a performance evaluation of a commercial auto-segmentation software on image series generated by a DECT.
View Article and Find Full Text PDFEur Radiol
December 2024
Department of Radiology, Antoni van Leeuwenhoek-The Netherlands Cancer Institute, Amsterdam, The Netherlands.
Objective: This study aims to assess and compare two state-of-the-art deep learning approaches for segmenting four thoracic organs at risk (OAR)-the esophagus, trachea, heart, and aorta-in CT images in the context of radiotherapy planning.
Materials And Methods: We compare a multi-organ segmentation approach and the fusion of multiple single-organ models, each dedicated to one OAR. All were trained using nnU-Net with the default parameters and the full-resolution configuration.
Comput Biol Med
December 2024
Aerospace Hi-tech Holding Group Co., LTD, Harbin, Heilongjiang, 150060, China.
CNN-based techniques have achieved impressive outcomes in medical image segmentation but struggle to capture long-term dependencies between pixels. The Transformer, with its strong feature extraction and representation learning abilities, performs exceptionally well within the domain of medical image partitioning. However, there are still shortcomings in bridging local to global connections, resulting in occasional loss of positional information.
View Article and Find Full Text PDFJ Cell Mol Med
December 2024
School of Computer Science and Software Engineering, University of Science and Technology Liaoning, Anshan, China.
Convolutional neural networks (CNNs) are well established in handling local features in visual tasks; yet, they falter in managing complex spatial relationships and long-range dependencies that are crucial for medical image segmentation, particularly in identifying pathological changes. While vision transformer (ViT) excels in addressing long-range dependencies, their ability to leverage local features remains inadequate. Recent ViT variants have merged CNNs to improve feature representation and segmentation outcomes, yet challenges with limited receptive fields and precise feature representation persist.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!