Purpose: The purpose of this study is to develop a Vision Transformer model with multitask classification framework that is appropriate for predicting four molecular expressions of glioma simultaneously based on MR imaging.
Materials And Methods: A total of 188 glioma (grades II-IV) patients with an immunohistochemical diagnosis of IDH, MGMT, Ki67 and P53 expression were enrolled in our study. A Vision Transformer (ViT) model, including three independent networks based on T2WI, T1CWI and T2 + T1CWI (T2-net, T1C-net and TU-net), was developed for the prediction of four glioma molecular expressions simultaneously. To evaluate the model performance, the accuracy rate, recall, precision, F1-score, and area under the receiver operating characteristic curve (AUC) were calculated.
Results: The proposed ViT model achieved high accuracy in predicting IDH, MGMT, Ki67 and P53 expression in gliomas. Among the three networks using the ViT model, TU-net achieved the best results with the highest values of accuracy (range, 0.937-0.969), precision (range, 0.949-0.972), recall (range, 0.873-0.991), F1-score (range, 0.910-0.981) and AUC (range, 0.976-0.984). Comparisons were also made between our ViT model and convolutional neural network (CNN)-based models, and the proposed ViT model outperformed the existing CNN-based models.
Conclusion: Vision Transformer is a reliable approach for the prediction of glioma molecular biomarkers and can be a viable alternative to CNNs.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1016/j.ejrad.2022.110560 | DOI Listing |
BMC Geriatr
January 2025
Department of Creative Product Design, Asia University, Taichung, Taiwan.
Alzheimer's disease (AD) is a complex, progressive, and irreversible neurodegenerative disorder marked by cognitive decline and memory loss. Early diagnosis is the most effective strategy to slow the disease's progression. Mild Cognitive Impairment (MCI) is frequently viewed as a crucial stage before the onset of AD, making it the ideal period for therapeutic intervention.
View Article and Find Full Text PDFSci Rep
January 2025
Electrical and Computer Engineering Department, University of Memphis, Memphis, TN, 38152, USA.
Oral squamous cell carcinoma (OSCC) is the most common form of oral cancer, with increasing global incidence and have poor prognosis. Tumour-infiltrating lymphocytes (TILs) are recognized as a key prognostic indicator and play a vital role in OSCC grading. However, current methods for TILs quantification are based on subjective visual assessments, leading to inter-observer variability and inconsistent diagnostic reproducibility.
View Article and Find Full Text PDFComput Biol Med
January 2025
Department of Biomedical Engineering, Islamic University, Kushtia, 7003, Bangladesh; Bio-Imaging Research Laboratory, Islamic University, Kushtia, 7003, Bangladesh. Electronic address:
Computed tomography (CT) scans play a key role in the diagnosis of stroke, a leading cause of morbidity and mortality worldwide. However, interpreting these scans is often challenging, necessitating automated solutions for timely and accurate diagnosis. This research proposed a novel hybrid model that integrates a Vision Transformer (ViT) and a Long Short Term Memory (LSTM) to accurately detect and classify stroke characteristics using CT images.
View Article and Find Full Text PDFTransl Vis Sci Technol
January 2025
Department of Biomedical Engineering, Faculty of Engineering, Mahidol University, Nakhon Pathom, Thailand.
Purpose: The purpose of this study was to develop a deep learning approach that restores artifact-laden optical coherence tomography (OCT) scans and predicts functional loss on the 24-2 Humphrey Visual Field (HVF) test.
Methods: This cross-sectional, retrospective study used 1674 visual field (VF)-OCT pairs from 951 eyes for training and 429 pairs from 345 eyes for testing. Peripapillary retinal nerve fiber layer (RNFL) thickness map artifacts were corrected using a generative diffusion model.
Sci Rep
January 2025
Affiliated Hospital 6 of Nantong University, Yancheng Third People's Hospital, Yancheng, 224001, Jiangsu, China.
Convolutional Neural Networks (CNNs) have achieved remarkable segmentation accuracy in medical image segmentation tasks. However, the Vision Transformer (ViT) model, with its capability of extracting global information, offers a significant advantage in contextual information compared to the limited receptive field of convolutional kernels in CNNs. Despite this, ViT models struggle to fully detect and extract high-frequency signals, such as textures and boundaries, in medical images.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!