Accuracy of visual cervical screening: verification bias revisited.

BJOG

Cancer Prevention Fellowship Program, Division of Cancer Prevention, National Cancer Institute, Rockville, MD, USA.

Published: April 2018

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC5742307PMC
http://dx.doi.org/10.1111/1471-0528.14797DOI Listing

Publication Analysis

Top Keywords

accuracy visual
4
visual cervical
4
cervical screening
4
screening verification
4
verification bias
4
bias revisited
4
accuracy
1
cervical
1
screening
1
verification
1

Similar Publications

Purpose: To develop and validate a prostate-specific membrane antigen (PSMA) PET/CT based multimodal deep learning model for predicting pathological lymph node invasion (LNI) in prostate cancer (PCa) patients identified as candidates for extended pelvic lymph node dissection (ePLND) by preoperative nomograms.

Methods: [Ga]Ga-PSMA-617 PET/CT scan of 116 eligible PCa patients (82 in the training cohort and 34 in the test cohort) who underwent radical prostatectomy with ePLND were analyzed in our study. The Med3D deep learning network was utilized to extract discriminative features from the entire prostate volume of interest on the PET/CT images.

View Article and Find Full Text PDF

In response to the pressing need for the detection of Monkeypox caused by the Monkeypox virus (MPXV), this study introduces the Enhanced Spatial-Awareness Capsule Network (ESACN), a Capsule Network architecture designed for the precise multi-class classification of dermatological images. Addressing the shortcomings of traditional Machine Learning and Deep Learning models, our ESACN model utilizes the dynamic routing and spatial hierarchy capabilities of CapsNets to differentiate complex patterns such as those seen in monkeypox, chickenpox, measles, and normal skin presentations. CapsNets' inherent ability to recognize and process crucial spatial relationships within images outperforms conventional CNNs, particularly in tasks that require the distinction of visually similar classes.

View Article and Find Full Text PDF

Optimizing autonomous artificial intelligence diagnostics for neuro-ocular health in space missions.

Life Sci Space Res (Amst)

February 2025

Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, NV, United States.

Spaceflight-Associated Neuro-Ocular Syndrome (SANS) presents a critical risk in long-duration missions, with microgravity-induced changes that threaten astronaut vision and mission outcomes. Current SANS monitoring, limited to pre- and post-flight exams, lacks in-flight diagnostics, highlighting an urgent need for autonomous tools capable of real-time assessment. Grok, an AI platform by xAI, offers promising potential as an advanced diagnostic tool for space-based health monitoring.

View Article and Find Full Text PDF

Background: The use of iodinated contrast-enhancing agents in computed tomography (CT) improves the visualization of relevant structures for radiotherapy treatment planning (RTP). However, it can lead to dose calculation errors by incorrectly converting a CT number to electron density.

Purpose: This study aimed to propose an algorithm for deriving virtual non-contrast (VNC) electron density from dual-energy CT (DECT) data.

View Article and Find Full Text PDF

In human-computer interaction, gesture recognition based on physiological signals offers advantages such as a more natural and fast interaction mode and less constrained by the environment than visual-based. Surface electromyography-based gesture recognition has significantly progressed. However, since individuals have physical differences, researchers must collect data multiple times from each user to train the deep learning model.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!