Accurate segmentation of organs or lesions from medical images is essential for accurate disease diagnosis and organ morphometrics. Previously, most researchers mainly added feature extraction modules and simply aggregated the semantic features to U-Net network to improve the segmentation accuracy of medical images. However, these improved U-Net networks ignore the semantic differences of different organs in medical images and lack the fusion of high-level semantic features and low-level semantic features, which will lead to blurred or miss boundaries between similar organs and diseased areas. To solve this problem, we propose Dual-branch dynamic hierarchical U-Net with multi-layer space fusion attention (D2HU-Net). Firstly, we propose a multi-layer spatial attention fusion module, which makes the shallow decoding path provide predictive graph supplement to the deep decoding path. Under the guidance of higher semantic features, useful context features are selected from lower semantic features to obtain deeper useful spatial information, which makes up for the semantic differences between organs in different medical images. Secondly, we propose a dynamic multi-scale layered module that enhances the multi-scale representation of the network at a finer granularity level and selectively refines single-scale features. Finally, the network provides guiding optimization for subsequent decoding based on multi-scale loss functions. The experimental results on four medical data sets show D2HU-Net enables the most advanced segmentation capabilities on different medical image datasets, which can help doctors diagnose and treat diseases.

Download full-text PDF

Source
http://dx.doi.org/10.1038/s41598-025-92715-0DOI Listing

Publication Analysis

Top Keywords

semantic features
20
medical images
16
dual-branch dynamic
8
dynamic hierarchical
8
hierarchical u-net
8
u-net multi-layer
8
multi-layer space
8
space fusion
8
fusion attention
8
medical image
8

Similar Publications

Accurate segmentation of organs or lesions from medical images is essential for accurate disease diagnosis and organ morphometrics. Previously, most researchers mainly added feature extraction modules and simply aggregated the semantic features to U-Net network to improve the segmentation accuracy of medical images. However, these improved U-Net networks ignore the semantic differences of different organs in medical images and lack the fusion of high-level semantic features and low-level semantic features, which will lead to blurred or miss boundaries between similar organs and diseased areas.

View Article and Find Full Text PDF

SAMP-Net: a medical image segmentation network with split attention and multi-layer perceptron.

Med Biol Eng Comput

March 2025

School of Electrical and Information Engineering, Beijing University of Civil Engineering and Architecture, No.15, Yongyuan Road, Huangcun Town, Daxing District, Beijing, 102616, China.

Convolutional neural networks (CNNs) have achieved remarkable success in computer vision, particularly in medical image segmentation. U-Net, a prominent architecture, marked a major breakthrough and remains widely used in practice. However, its uniform downsampling strategy and simple stacking of convolutional layers in the encoder limit its ability to capture rich features at multiple depths, reducing its efficiency for rapid image processing.

View Article and Find Full Text PDF

Recently, contrast-enhanced ultrasound (CEUS) has presented a potential value in the diagnosis of liver trauma, the leading cause of death in blunt abdominal trauma. However, the inherent speckle noise and the complicated visual characteristics of blunt liver trauma in CEUS images make the diagnosis highly dependent on the expertise of radiologists, which is subjective and time-consuming. Moreover, the intra- and inter-observer variance inevitably influences the accuracy of diagnosis using CEUS.

View Article and Find Full Text PDF

Image quality assessment (IQA) plays a critical role in automatically detecting and correcting defects in images, thereby enhancing the overall performance of image processing and transmission systems. While research on reference-based IQA is well-established, studies on no-reference image IQA remain underdeveloped. In this article, we propose a novel no-reference IQA algorithm based on transfer learning (IQA-NRTL).

View Article and Find Full Text PDF

Novel cross-dimensional coarse-fine-grained complementary network for image-text matching.

PeerJ Comput Sci

March 2025

Department of Electrical Engineering, Faculty of Engineering, Universiti Malaya, Kuala Lumpur, Malaysia.

The fundamental aspects of multimodal applications such as image-text matching, and cross-modal heterogeneity gap between images and texts have always been challenging and complex. Researchers strive to overcome the challenges by proposing numerous significant efforts directed toward narrowing the semantic gap between visual and textual modalities. However, existing methods are usually limited to computing the similarity between images (image regions) and text (text words), ignoring the semantic consistency between fine-grained matching of word regions and coarse-grained overall matching of image and text.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!