Medical visual question answering (VQA) aims to correctly answer a clinical question related to a given medical image. Nevertheless, owing to the expensive manual annotations of medical data, the lack of labeled data limits the development of medical VQA. In this paper, we propose a simple yet effective data augmentation method, VQAMix, to mitigate the data limitation problem. Specifically, VQAMix generates more labeled training samples by linearly combining a pair of VQA samples, which can be easily embedded into any visual-language model to boost performance. However, mixing two VQA samples would construct new connections between images and questions from different samples, which will cause the answers for those new fabricated image-question pairs to be missing or meaningless. To solve the missing answer problem, we first develop the Learning with Missing Labels (LML) strategy, which roughly excludes the missing answers. To alleviate the meaningless answer issue, we design the Learning with Conditional-mixed Labels (LCL) strategy, which further utilizes language-type prior to forcing the mixed pairs to have reasonable answers that belong to the same category. Experimental results on the VQA-RAD and PathVQA benchmarks show that our proposed method significantly improves the performance of the baseline by about 7% and 5% on the averaging result of two backbones, respectively. More importantly, VQAMix could improve confidence calibration and model interpretability, which is significant for medical VQA models in practical applications. All code and models are available at https://github.com/haifangong/VQAMix.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TMI.2022.3185008DOI Listing

Publication Analysis

Top Keywords

medical visual
8
visual question
8
question answering
8
medical vqa
8
vqa samples
8
medical
6
vqa
5
vqamix
4
vqamix conditional
4
conditional triplet
4

Similar Publications

Background: Clinical decision support systems leveraging artificial intelligence (AI) are increasingly integrated into health care practices, including pharmacy medication verification. Communicating uncertainty in an AI prediction is viewed as an important mechanism for boosting human collaboration and trust. Yet, little is known about the effects on human cognition as a result of interacting with such types of AI advice.

View Article and Find Full Text PDF

Objective: Although the 5-level EQ-5D version (EQ-5D-5L) instrument has been used to determine health-related quality of life and health utility in gout, it is used in comparing health utility among gout flare (GF) and non-gout flare (non-GF) patients is still limited. This study aimed to compare health utility among GF and non-GF patients in Thailand.

Methods: In this multi-center cross-sectional study, patients with GF and non-GF were interviewed for the EQ-5D5L and EQ-Visual Analog Scale (VAS) instruments by rheumatologists or trained research staffs.

View Article and Find Full Text PDF

Purpose: To investigate the presence of tubulation in the outer nuclear layer of patients with Bietti crystalline dystrophy (BCD) using optical coherence tomography (OCT) and evaluate its relationship with visual field, visual field progression, and retinal volume.

Methods: This retrospective cross-sectional study included 37 patients diagnosed with BCD who underwent spectral-domain OCT examination. OCT examinations and Humphrey visual field tests (10-2 program) were conducted.

View Article and Find Full Text PDF

Purpose: Low-dose CT (LDCT) screening effectively reduces lung adenocarcinoma (LUAD) mortality. However, accurately evaluating the malignant potential of indeterminate lung nodules remains a challenge. Carcinoembryonic antigen cell adhesion molecule 6 (CEACAM6), a potential biomarker for distinguishing benign pulmonary nodules from LUAD, may be leveraged for noninvasive positron emission tomography (PET) imaging to aid LUAD diagnosis.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!