Publications by authors named "Jong Hak Moon"

Recently a number of studies demonstrated impressive performance on diverse vision-language multi-modal tasks such as image captioning and visual question answering by extending the BERT architecture with multi-modal pre-training objectives. In this work we explore a broad set of multi-modal representation learning tasks in the medical domain, specifically using radiology images and the unstructured report. We propose Medical Vision Language Learner (MedViLL), which adopts a BERT-based architecture combined with a novel multi-modal attention masking scheme to maximize generalization performance for both vision-language understanding tasks (diagnosis classification, medical image-report retrieval, medical visual question answering) and vision-language generation task (radiology report generation).

View Article and Find Full Text PDF
Article Synopsis
  • Coronary artery disease is a major cause of death, and the standard evaluation method—coronary angiography—often faces challenges due to variability in readings, prompting the need for automated solutions.
  • A deep-learning algorithm has been developed to automatically detect and classify stenosis (narrowing of arteries) in coronary angiographic images, utilizing key frame extraction and a self-attention mechanism for improved accuracy.
  • The model demonstrated impressive results, achieving high accuracy in both internal and external validations and effectively visualizing stenosis locations through advanced techniques like gradient-weighted class activation mapping.*
View Article and Find Full Text PDF

Background: It is necessary to consider myopic optic disc tilt as it seriously impacts normal ocular parameters. However, ophthalmologic measurements are within inter-observer variability and time-consuming to get. This study aimed to develop and evaluate deep learning models that automatically recognize a myopic tilted optic disc in fundus photography.

View Article and Find Full Text PDF