Objective: This study aimed to develop an open-source multimodal large language model (CXR-LLaVA) for interpreting chest X-ray images (CXRs), leveraging recent advances in large language models (LLMs) to potentially replicate the image interpretation skills of human radiologists.
Materials And Methods: For training, we collected 592,580 publicly available CXRs, of which 374,881 had labels for certain radiographic abnormalities (Dataset 1) and 217,699 provided free-text radiology reports (Dataset 2). After pre-training a vision transformer with Dataset 1, we integrated it with an LLM influenced by the LLaVA network. Then, the model was fine-tuned, primarily using Dataset 2. The model's diagnostic performance for major pathological findings was evaluated, along with the acceptability of radiologic reports by human radiologists, to gauge its potential for autonomous reporting.
Results: The model demonstrated impressive performance in test sets, achieving an average F1 score of 0.81 for six major pathological findings in the MIMIC internal test set and 0.56 for six major pathological findings in the external test set. The model's F1 scores surpassed those of GPT-4-vision and Gemini-Pro-Vision in both test sets. In human radiologist evaluations of the external test set, the model achieved a 72.7% success rate in autonomous reporting, slightly below the 84.0% rate of ground truth reports.
Conclusion: This study highlights the significant potential of multimodal LLMs for CXR interpretation, while also acknowledging the performance limitations. Despite these challenges, we believe that making our model open-source will catalyze further research, expanding its effectiveness and applicability in various clinical contexts.
Key Points: Question How can a multimodal large language model be adapted to interpret chest X-rays and generate radiologic reports? Findings The developed CXR-LLaVA model effectively detects major pathological findings in chest X-rays and generates radiologic reports with a higher accuracy compared to general-purpose models. Clinical relevance This study demonstrates the potential of multimodal large language models to support radiologists by autonomously generating chest X-ray reports, potentially reducing diagnostic workloads and improving radiologist efficiency.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1007/s00330-024-11339-6 | DOI Listing |
JAMA Netw Open
January 2025
Laboratory of NeuroImaging, National Institute on Alcohol Abuse and Alcoholism, Bethesda, Maryland.
Importance: Cannabis use has increased globally, but its effects on brain function are not fully known, highlighting the need to better determine recent and long-term brain activation outcomes of cannabis use.
Objective: To examine the association of lifetime history of heavy cannabis use and recent cannabis use with brain activation across a range of brain functions in a large sample of young adults in the US.
Design, Setting, And Participants: This cross-sectional study used data (2017 release) from the Human Connectome Project (collected between August 2012 and 2015).
Surg Radiol Anat
January 2025
Department of Neurosurgery, Saitama Sekishinkai Hospital, 2-37-20 Irumagawa, Sayama, Saitama, 350-1305, Japan.
Purpose: To describe a case of short common trunk of the occipital artery (OA) and ascending pharyngeal artery (APA) arising from the internal carotid artery (ICA).
Methods: A 36-year-old woman with a history of surgical resection of a right lateral ventricular meningioma and atheromatous plaque of the right ICA underwent cranial magnetic resonance (MR) imaging and MR angiography of the head and neck region with a 3-Tesla scanner.
Results: MR angiography of the neck region showed a small atheromatous plaque at the origin of the right ICA and an anomalous artery arising from the posteromedial aspect of the right ICA at the distal end of the carotid bulb.
Radiology
January 2025
From the Departments of Biomedical Systems Informatics (S.K., Jaewoong Kim, C.H., D.Y.) and Neurology (Joonho Kim, J.Y.), Yonsei University College of Medicine, 50-1 Yonsei-ro, Seodaemun-gu, Seoul 03722, Republic of Korea; Department of Radiology, Central Draft Physical Examination Office of Military Manpower Administration, Daegu, Republic of Korea (D.K.); Department of Radiology, Research Institute of Radiological Science and Center for Clinical Imaging Data Science (H.J.S. Y.K., S.J.), and Center for Digital Health (H.J.S., D.Y.), Yongin Severance Hospital, Yonsei University College of Medicine, Yongin, Republic of Korea; Department of Radiology, Gangnam Severance Hospital, Yonsei University College of Medicine, Seoul, Republic of Korea (S.H.L.); Departments of Radiology (M.H.) and Neurology (S.J.L.), Ajou University Hospital, Ajou University School of Medicine, Suwon, Republic of Korea; and Institute for Innovation in Digital Healthcare, Severance Hospital, Seoul, Republic of Korea (D.Y.).
Background The increasing workload of radiologists can lead to burnout and errors in radiology reports. Large language models, such as OpenAI's GPT-4, hold promise as error revision tools for radiology. Purpose To test the feasibility of GPT-4 use by determining its error detection, reasoning, and revision performance on head CT reports with varying error types and to validate its clinical utility by comparison with human readers.
View Article and Find Full Text PDFRadiology
January 2025
From the University of Maryland Medical Intelligent Imaging (UM2ii) Center, Department of Diagnostic Radiology and Nuclear Medicine, University of Maryland School of Medicine, 22 S Greene St, Baltimore, MD 21201 (C.H.S., A.K., V.P., F.X.D.); Departments of Radiology, Medicine, and Biomedical Data Science, Stanford University, Palo Alto, Calif (C.P.L.); Department of Computer Science and Electrical Engineering, College of Engineering and Information Technology, University of Maryland, Baltimore County, Baltimore, Md (A.J.); Department of Computer Science, University of Maryland, College Park, College Park, Md (H.H.); and University of Maryland Institute for Health Computing, University of Maryland, North Bethesda, Md (H.H., F.X.D.).
Integrating large language models (LLMs) into health care holds substantial potential to enhance clinical workflows and care delivery. However, LLMs also pose serious risks if integration is not thoughtfully executed, with complex challenges spanning accuracy, accessibility, privacy, and regulation. Proprietary commercial LLMs (eg, GPT-4 [OpenAI], Claude 3 Sonnet and Claude 3 Opus [Anthropic], Gemini [Google]) have received much attention from researchers in the medical domain, including radiology.
View Article and Find Full Text PDFJ Deaf Stud Deaf Educ
January 2025
Institute of Neurology of Senses and Language, Hospital of St. John of God Linz, Linz, Austria.
Language comprehension is an essential component of human development that is associated not only with expressive language development and knowledge acquisition, but also with social inclusion, mental health, and quality of life. For deaf and hard-of-hearing adults with intellectual disability, there is a paucity of measures of receptive sign language skills, although these are a prerequisite for individualized planning and evaluation of intervention. Assessments require materials and procedures that are accurate, feasible, and suitable for low levels of functioning.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!