Hematoxylin and Eosin (H&E) staining is a widely used sample preparation procedure for enhancing the saturation of tissue sections and the contrast between nuclei and cytoplasm in histology images for medical diagnostics. However, various factors, such as the differences in the reagents used, result in high variability in the colors of the stains actually recorded. This variability poses a challenge in achieving generalization for machine-learning based computer-aided diagnostic tools. To desensitize the learned models to stain variations, we propose the Generative Stain Augmentation Network (G-SAN) - a GAN-based framework that augments a collection of cell images with simulated yet realistic stain variations. At its core, G-SAN uses a novel and highly computationally efficient Laplacian Pyramid (LP) based generator architecture, that is capable of disentangling stain from cell morphology. Through the task of patch classification and nucleus segmentation, we show that using G-SAN-augmented training data provides on average 15.7% improvement in F1 score and 7.3% improvement in panoptic quality, respectively. Our code is available at https://github.com/lifangda01/GSAN-Demo.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TMI.2023.3317239 | DOI Listing |
Sci Rep
January 2025
Affiliated Hospital 6 of Nantong University, Yancheng Third People's Hospital, Yancheng, 224001, Jiangsu, China.
Convolutional Neural Networks (CNNs) have achieved remarkable segmentation accuracy in medical image segmentation tasks. However, the Vision Transformer (ViT) model, with its capability of extracting global information, offers a significant advantage in contextual information compared to the limited receptive field of convolutional kernels in CNNs. Despite this, ViT models struggle to fully detect and extract high-frequency signals, such as textures and boundaries, in medical images.
View Article and Find Full Text PDFComput Methods Programs Biomed
November 2024
East China University of Science and Technology, Shanghai, China. Electronic address:
Background And Objective: Accurate medical image segmentation is crucial for diagnosis and treatment planning, particularly in tumor localization and organ measurement. Despite the success of Transformer models in various domains, they still struggle to capture high-frequency features, limiting their performance in medical image segmentation, especially in edge texture extraction. To overcome this limitation and improve segmentation accuracy, this study proposes a novel model architecture aimed at enhancing the Transformer's ability to capture and integrate both high-frequency and low-frequency features.
View Article and Find Full Text PDFSci Rep
November 2024
College of Material Science and Art Design, Inner Mongolia Agricultural University, Hohhot, 010018, China.
Aiming at the veneer defect image acquisition process is prone to the problems of blurred edges, inconspicuous contrast and distortion, which cannot show the defects clearly.To improve image analyzability and clarity, a veneer defect image enhancement method based on AMEF-AGC is proposed herein. First, a veneer defect image is subjected to Gamma correction to obtain multiple underexposed image sequences for which Gaussian and Laplacian pyramids are constructed to determine the weights of the multiple exposure sequence group images.
View Article and Find Full Text PDFSci Rep
October 2024
College of Computer and Communication, Hunan Institute of Engineering, Xiangtan, 411104, Hunan, China.
Monocular depth estimation is an important but challenging task. Although the performance has been improved by adopting various encoder-decoder architectures, the estimated depth maps lack structure details and clear edges due to simple repeated upsampling. To solve this problem, this paper presents the novel LapUNet (Laplacian U-shape networks), in which the encoder adopts ResNeXt101, and the decoder is constructed with the novel DLRU (dynamic Laplacian residual U-shape) module.
View Article and Find Full Text PDFIEEE Trans Image Process
October 2024
Deep learning-based video denoising methods have achieved great performance improvements in recent years. However, the expensive computational cost arising from sophisticated network design has severely limited their applications in real-world scenarios. To address this practical weakness, we propose a multiscale spatio-temporal memory network for fast video denoising, named MSTMN, aiming at striking an improved trade-off between cost and performance.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!