Hematoxylin and Eosin (H&E) staining is a widely used sample preparation procedure for enhancing the saturation of tissue sections and the contrast between nuclei and cytoplasm in histology images for medical diagnostics. However, various factors, such as the differences in the reagents used, result in high variability in the colors of the stains actually recorded. This variability poses a challenge in achieving generalization for machine-learning based computer-aided diagnostic tools. To desensitize the learned models to stain variations, we propose the Generative Stain Augmentation Network (G-SAN) - a GAN-based framework that augments a collection of cell images with simulated yet realistic stain variations. At its core, G-SAN uses a novel and highly computationally efficient Laplacian Pyramid (LP) based generator architecture, that is capable of disentangling stain from cell morphology. Through the task of patch classification and nucleus segmentation, we show that using G-SAN-augmented training data provides on average 15.7% improvement in F1 score and 7.3% improvement in panoptic quality, respectively. Our code is available at https://github.com/lifangda01/GSAN-Demo.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TMI.2023.3317239DOI Listing

Publication Analysis

Top Keywords

laplacian pyramid
8
pyramid based
8
stain augmentation
8
augmentation network
8
stain variations
8
stain
5
based generative
4
generative h&e
4
h&e stain
4
network hematoxylin
4

Similar Publications

Convolutional Neural Networks (CNNs) have achieved remarkable segmentation accuracy in medical image segmentation tasks. However, the Vision Transformer (ViT) model, with its capability of extracting global information, offers a significant advantage in contextual information compared to the limited receptive field of convolutional kernels in CNNs. Despite this, ViT models struggle to fully detect and extract high-frequency signals, such as textures and boundaries, in medical images.

View Article and Find Full Text PDF

Background And Objective: Accurate medical image segmentation is crucial for diagnosis and treatment planning, particularly in tumor localization and organ measurement. Despite the success of Transformer models in various domains, they still struggle to capture high-frequency features, limiting their performance in medical image segmentation, especially in edge texture extraction. To overcome this limitation and improve segmentation accuracy, this study proposes a novel model architecture aimed at enhancing the Transformer's ability to capture and integrate both high-frequency and low-frequency features.

View Article and Find Full Text PDF

Aiming at the veneer defect image acquisition process is prone to the problems of blurred edges, inconspicuous contrast and distortion, which cannot show the defects clearly.To improve image analyzability and clarity, a veneer defect image enhancement method based on AMEF-AGC is proposed herein. First, a veneer defect image is subjected to Gamma correction to obtain multiple underexposed image sequences for which Gaussian and Laplacian pyramids are constructed to determine the weights of the multiple exposure sequence group images.

View Article and Find Full Text PDF

Monocular depth estimation is an important but challenging task. Although the performance has been improved by adopting various encoder-decoder architectures, the estimated depth maps lack structure details and clear edges due to simple repeated upsampling. To solve this problem, this paper presents the novel LapUNet (Laplacian U-shape networks), in which the encoder adopts ResNeXt101, and the decoder is constructed with the novel DLRU (dynamic Laplacian residual U-shape) module.

View Article and Find Full Text PDF

Deep learning-based video denoising methods have achieved great performance improvements in recent years. However, the expensive computational cost arising from sophisticated network design has severely limited their applications in real-world scenarios. To address this practical weakness, we propose a multiscale spatio-temporal memory network for fast video denoising, named MSTMN, aiming at striking an improved trade-off between cost and performance.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!