Purpose: The purpose of this study is to develop a Vision Transformer model with multitask classification framework that is appropriate for predicting four molecular expressions of glioma simultaneously based on MR imaging.

Materials And Methods: A total of 188 glioma (grades II-IV) patients with an immunohistochemical diagnosis of IDH, MGMT, Ki67 and P53 expression were enrolled in our study. A Vision Transformer (ViT) model, including three independent networks based on T2WI, T1CWI and T2 + T1CWI (T2-net, T1C-net and TU-net), was developed for the prediction of four glioma molecular expressions simultaneously. To evaluate the model performance, the accuracy rate, recall, precision, F1-score, and area under the receiver operating characteristic curve (AUC) were calculated.

Results: The proposed ViT model achieved high accuracy in predicting IDH, MGMT, Ki67 and P53 expression in gliomas. Among the three networks using the ViT model, TU-net achieved the best results with the highest values of accuracy (range, 0.937-0.969), precision (range, 0.949-0.972), recall (range, 0.873-0.991), F1-score (range, 0.910-0.981) and AUC (range, 0.976-0.984). Comparisons were also made between our ViT model and convolutional neural network (CNN)-based models, and the proposed ViT model outperformed the existing CNN-based models.

Conclusion: Vision Transformer is a reliable approach for the prediction of glioma molecular biomarkers and can be a viable alternative to CNNs.

Download full-text PDF

Source
http://dx.doi.org/10.1016/j.ejrad.2022.110560DOI Listing

Publication Analysis

Top Keywords

vit model
20
vision transformer
16
molecular expressions
12
multitask classification
8
classification framework
8
predicting molecular
8
expressions glioma
8
idh mgmt
8
mgmt ki67
8
ki67 p53
8

Similar Publications

Alzheimer's disease (AD) is a complex, progressive, and irreversible neurodegenerative disorder marked by cognitive decline and memory loss. Early diagnosis is the most effective strategy to slow the disease's progression. Mild Cognitive Impairment (MCI) is frequently viewed as a crucial stage before the onset of AD, making it the ideal period for therapeutic intervention.

View Article and Find Full Text PDF

Oral squamous cell carcinoma (OSCC) is the most common form of oral cancer, with increasing global incidence and have poor prognosis. Tumour-infiltrating lymphocytes (TILs) are recognized as a key prognostic indicator and play a vital role in OSCC grading. However, current methods for TILs quantification are based on subjective visual assessments, leading to inter-observer variability and inconsistent diagnostic reproducibility.

View Article and Find Full Text PDF

A novel hybrid ViT-LSTM model with explainable AI for brain stroke detection and classification in CT images: A case study of Rajshahi region.

Comput Biol Med

January 2025

Department of Biomedical Engineering, Islamic University, Kushtia, 7003, Bangladesh; Bio-Imaging Research Laboratory, Islamic University, Kushtia, 7003, Bangladesh. Electronic address:

Computed tomography (CT) scans play a key role in the diagnosis of stroke, a leading cause of morbidity and mortality worldwide. However, interpreting these scans is often challenging, necessitating automated solutions for timely and accurate diagnosis. This research proposed a novel hybrid model that integrates a Vision Transformer (ViT) and a Long Short Term Memory (LSTM) to accurately detect and classify stroke characteristics using CT images.

View Article and Find Full Text PDF

Purpose: The purpose of this study was to develop a deep learning approach that restores artifact-laden optical coherence tomography (OCT) scans and predicts functional loss on the 24-2 Humphrey Visual Field (HVF) test.

Methods: This cross-sectional, retrospective study used 1674 visual field (VF)-OCT pairs from 951 eyes for training and 429 pairs from 345 eyes for testing. Peripapillary retinal nerve fiber layer (RNFL) thickness map artifacts were corrected using a generative diffusion model.

View Article and Find Full Text PDF

Convolutional Neural Networks (CNNs) have achieved remarkable segmentation accuracy in medical image segmentation tasks. However, the Vision Transformer (ViT) model, with its capability of extracting global information, offers a significant advantage in contextual information compared to the limited receptive field of convolutional kernels in CNNs. Despite this, ViT models struggle to fully detect and extract high-frequency signals, such as textures and boundaries, in medical images.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!