AGSAM: Agent-Guided Segment Anything Model for Automatic Segmentation in Few-Shot Scenarios.

Bioengineering (Basel)

State Key Laboratory of Ophthalmology, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Zhongshan Ophthalmic Center, Sun Yat-Sen University, Guangzhou 510000, China.

Published: April 2024

Precise medical image segmentation of regions of interest (ROIs) is crucial for accurate disease diagnosis and progression assessment. However, acquiring high-quality annotated data at the pixel level poses a significant challenge due to the resource-intensive nature of this process. This scarcity of high-quality annotated data results in few-shot scenarios, which are highly prevalent in clinical applications. To address this obstacle, this paper introduces Agent-Guided SAM (AGSAM), an innovative approach that transforms the Segment Anything Model (SAM) into a fully automated segmentation method by automating prompt generation. Capitalizing on the pre-trained feature extraction and decoding capabilities of SAM-Med2D, AGSAM circumvents the need for manual prompt engineering, ensuring adaptability across diverse segmentation methods. Furthermore, the proposed feature augmentation convolution module (FACM) enhances model accuracy by promoting stable feature representations. Experimental evaluations demonstrate AGSAM's consistent superiority over other methods across various metrics. These findings highlight AGSAM's efficacy in tackling the challenges associated with limited annotated data while achieving high-quality medical image segmentation.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11118214PMC
http://dx.doi.org/10.3390/bioengineering11050447DOI Listing

Publication Analysis

Top Keywords

annotated data
12
segment model
8
few-shot scenarios
8
medical image
8
image segmentation
8
high-quality annotated
8
segmentation
5
agsam agent-guided
4
agent-guided segment
4
model automatic
4

Similar Publications

In this comprehensive review, we delve into the transformative role of artificial intelligence (AI) in refining the application of multi-omics and spatial multi-omics within the realm of diffuse large B-cell lymphoma (DLBCL) research. We scrutinized the current landscape of multi-omics and spatial multi-omics technologies, accentuating their combined potential with AI to provide unparalleled insights into the molecular intricacies and spatial heterogeneity inherent to DLBCL. Despite current progress, we acknowledge the hurdles that impede the full utilization of these technologies, such as the integration and sophisticated analysis of complex datasets, the necessity for standardized protocols, the reproducibility of findings, and the interpretation of their biological significance.

View Article and Find Full Text PDF

Objective: To improve performance of medical entity normalization across many languages, especially when fewer language resources are available compared to English.

Materials And Methods: We propose xMEN, a modular system for cross-lingual (x) medical entity normalization (MEN), accommodating both low- and high-resource scenarios. To account for the scarcity of aliases for many target languages and terminologies, we leverage multilingual aliases via cross-lingual candidate generation.

View Article and Find Full Text PDF

Deep learning-based automated diagnosis of temporomandibular joint anterior disc displacement and its clinical application.

Front Physiol

December 2024

Department of Oral & Maxillofacial Surgery, Shenzhen Stomatology Hospital, Affiliated to Shenzhen University, Shenzhen, Guangdong Province, China.

Introduction: This study aimed to develop a deep learning-based method for interpreting magnetic resonance imaging (MRI) scans of temporomandibular joint (TMJ) anterior disc displacement (ADD) and to formulate an automated diagnostic system for clinical practice.

Methods: The deep learning models were utilized to identify regions of interest (ROI), segment TMJ structures including the articular disc, condyle, glenoid fossa, and articular tubercle, and classify TMJ ADD. The models employed Grad-CAM heatmaps and segmentation annotation diagrams for visual diagnostic predictions and were deployed for clinical application.

View Article and Find Full Text PDF

Background: Pneumothorax is a medical emergency caused by the abnormal accumulation of air in the pleural space-the potential space between the lungs and chest wall. On 2D chest radiographs, pneumothorax occurs within the thoracic cavity and outside of the mediastinum, and we refer to this area as "lung + space." While deep learning (DL) has increasingly been utilized to segment pneumothorax lesions in chest radiographs, many existing DL models employ an end-to-end approach.

View Article and Find Full Text PDF

A novel methodology for dataset augmentation in the semantic segmentation of coil-coated surface degradation is presented in this study. Deep convolutional generative adversarial networks (DCGAN) are employed to generate synthetic input-target pairs, which closely resemble real-world data, with the goal of expanding an existing dataset. These augmented datasets are used to train two state-of-the-art models, U-net, and DeepLabV3, for the precise detection of degradation areas around scribes.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!