Abnormality-aware multimodal learning for WSI classification.

Front Med (Lausanne)

Department of Computer Science and Engineering, University of Texas at Arlington, Arlington, TX, United States.

Published: February 2025

Whole slide images (WSIs) play a vital role in cancer diagnosis and prognosis. However, their gigapixel resolution, lack of pixel-level annotations, and reliance on unimodal visual data present challenges for accurate and efficient computational analysis. Existing methods typically divide WSIs into thousands of patches, which increases computational demands and makes it challenging to effectively focus on diagnostically relevant regions. Furthermore, these methods frequently rely on feature extractors pretrained on natural images, which are not optimized for pathology tasks, and overlook multimodal data sources such as cellular and textual information that can provide critical insights. To address these limitations, we propose the bnormality-ware ultiodal (AAMM) learning framework, which integrates abnormality detection and multimodal feature learning for WSI classification. AAMM incorporates a Gaussian Mixture Variational Autoencoder (GMVAE) to identify and select the most informative patches, reducing computational complexity while retaining critical diagnostic information. It further integrates multimodal features from pathology-specific foundation models, combining patch-level, cell-level, and text-level representations through cross-attention mechanisms. This approach enhances the ability to comprehensively analyze WSIs for cancer diagnosis and subtyping. Extensive experiments on normal-tumor classification and cancer subtyping demonstrate that AAMM achieves superior performance compared to state-of-the-art methods. By combining abnormal detection with multimodal feature integration, our framework offers an efficient and scalable solution for advancing computational pathology.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11893561PMC
http://dx.doi.org/10.3389/fmed.2025.1546452DOI Listing

Publication Analysis

Top Keywords

learning wsi
8
wsi classification
8
cancer diagnosis
8
detection multimodal
8
multimodal feature
8
abnormality-aware multimodal
4
multimodal learning
4
classification slide
4
slide images
4
images wsis
4

Similar Publications

Abnormality-aware multimodal learning for WSI classification.

Front Med (Lausanne)

February 2025

Department of Computer Science and Engineering, University of Texas at Arlington, Arlington, TX, United States.

Whole slide images (WSIs) play a vital role in cancer diagnosis and prognosis. However, their gigapixel resolution, lack of pixel-level annotations, and reliance on unimodal visual data present challenges for accurate and efficient computational analysis. Existing methods typically divide WSIs into thousands of patches, which increases computational demands and makes it challenging to effectively focus on diagnostically relevant regions.

View Article and Find Full Text PDF

A foundation model for generalizable cancer diagnosis and survival prediction from histopathological images.

Nat Commun

March 2025

Department of Bioinformatics and Biostatistics, School of Life Sciences and Biotechnology, Shanghai Jiao Tong University, Shanghai, China.

Computational pathology, utilizing whole slide images (WSIs) for pathological diagnosis, has advanced the development of intelligent healthcare. However, the scarcity of annotated data and histological differences hinder the general application of existing methods. Extensive histopathological data and the robustness of self-supervised models in small-scale data demonstrate promising prospects for developing foundation pathology models.

View Article and Find Full Text PDF

Digital pathology offers a groundbreaking opportunity to transform clinical practice in histopathological image analysis, yet faces a significant hurdle: the substantial file sizes of pathological whole slide images (WSIs). Whereas current digital pathology solutions rely on lossy JPEG compression to address this issue, lossy compression can introduce color and texture disparities, potentially impacting clinical decision-making. Whereas prior research addresses perceptual image quality and downstream performance independently of each other, we jointly evaluate compression schemes for perceptual and downstream task quality on four different datasets.

View Article and Find Full Text PDF

Development of a deep learning-based model to evaluate changes during radiotherapy using cervical cancer digital pathology.

J Radiat Res

March 2025

Department of Radiation Oncology & Proton Medical Research Center, Institute of Medicine, University of Tsukuba, 2-1-1 Amakubo, Tsubuka, Ibaraki 305-8576, Japan.

This study aims to create a deep learning-based classification model for cervical cancer biopsy before and during radiotherapy, visualize the results on whole slide images (WSIs), and explore the clinical significance of obtained features. This study included 95 patients with cervical cancer who received radiotherapy between April 2013 and December 2020. Hematoxylin-eosin stained biopsies were digitized to WSIs and divided into small tiles.

View Article and Find Full Text PDF

Whole slide image (WSI) classification plays a crucial role in digital pathology data analysis. However, the immense size of WSIs and the absence of fine-grained sub-region labels pose significant challenges for accurate WSI classification. Typical classification-driven deep learning methods often struggle to generate informative image representations, which can compromise the robustness of WSI classification.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!