Optimizing Large Language Models in Radiology and Mitigating Pitfalls: Prompt Engineering and Fine-tuning.

Radiographics

From the Department of Radiology, George Washington University School of Medicine and Health Sciences, 2300 I St NW, Washington, DC 20052 (T.T.K., R.J.); Yale School of Medicine, New Haven, Conn (M.M.); and University of California San Francisco, San Francisco, Calif (R.S.).

Published: April 2025

Large language models (LLMs) such as generative pretrained transformers (GPTs) have had a major impact on society, and there is increasing interest in using these models for applications in medicine and radiology. This article presents techniques to optimize these models and describes their known challenges and limitations. Specifically, the authors explore how to best craft natural language prompts, a process known as prompt engineering, for these models to elicit more accurate and desirable responses. The authors also explain how fine-tuning is conducted, in which a more general model, such as GPT-4, is further trained on a more specific use case, such as summarizing clinical notes, to further improve reliability and relevance. Despite the enormous potential of these models, substantial challenges limit their widespread implementation. These tools differ substantially from traditional health technology in their complexity and their probabilistic and nondeterministic nature, and these differences lead to issues such as "hallucinations," biases, lack of reliability, and security risks. Therefore, the authors provide radiologists with baseline knowledge of the technology underpinning these models and an understanding of how to use them, in addition to exploring best practices in prompt engineering and fine-tuning. Also discussed are current proof-of-concept use cases of LLMs in the radiology literature, such as in clinical decision support and report generation, and the limitations preventing their current adoption in medicine and radiology. RSNA, 2025 See invited commentary by Chung and Mongan in this issue.

Download full-text PDF

Source
http://dx.doi.org/10.1148/rg.240073DOI Listing

Publication Analysis

Top Keywords

prompt engineering
12
large language
8
language models
8
engineering fine-tuning
8
medicine radiology
8
models
7
optimizing large
4
radiology
4
models radiology
4
radiology mitigating
4

Similar Publications

FedBM: Stealing knowledge from pre-trained language models for heterogeneous federated learning.

Med Image Anal

March 2025

Department of Mechanical Engineering, City University of Hong Kong, Hong Kong Special Administrative Region of China; Department of Data and Systems Engineering, The University of Hong Kong, Hong Kong Special Administrative Region of China. Electronic address:

Federated learning (FL) has shown great potential in medical image computing since it provides a decentralized learning paradigm that allows multiple clients to train a model collaboratively without privacy leakage. However, current studies have shown that data heterogeneity incurs local learning bias in classifiers and feature extractors of client models during local training, leading to the performance degradation of a federation system. To address these issues, we propose a novel framework called Federated Bias eliMinating (FedBM) to get rid of local learning bias in heterogeneous federated learning (FL), which mainly consists of two modules, i.

View Article and Find Full Text PDF

Prompt-based polyp segmentation during endoscopy.

Med Image Anal

February 2025

School of Computing and Mathematical Sciences, University of Leicester, Leicester LE1 7RH, UK.

Accurate judgment and identification of polyp size is crucial in endoscopic diagnosis. However, the indistinct boundaries of polyps lead to missegmentation and missed cancer diagnoses. In this paper, a prompt-based polyp segmentation method (PPSM) is proposed to assist in early-stage cancer diagnosis during endoscopy.

View Article and Find Full Text PDF

In particle therapy (PT), several methods are being investigated to help reduce range margins and identify deviations from the original treatment plan, such as prompt-gamma (PG) imaging with Compton cameras (CC). To reconstruct the images, the Origin Ensemble (OE) algorithm is commonly used. In the context of PT, artifacts and strong noise often affect CC images.

View Article and Find Full Text PDF

Teleost IgM+ plasma-like cells: beyond antibody secretion.

J Immunol

January 2025

Biotechnology Department, Instituto Nacional de Investigación y Tecnología Agraria y Alimentaria, Consejo Superior de Investigaciones Científicas, Madrid, Spain.

Upon antigen encounter, B cells start a differentiation process toward antibody-secreting cells (ASCs), initially plasmablasts, and eventually long-lived plasma cells. All these ASCs specialize in secreting important amounts of antibodies and usually lose other functionalities of naïve B cells. This differentiation process is scarcely characterized in teleost fish, in which B cells have been shown to share many functional and phenotypic characteristics of mammalian B1 innate subsets.

View Article and Find Full Text PDF

Motivation: Inferring gene networks provides insights into biological pathways and functional relationships among genes. When gene expression samples exhibit heterogeneity, they may originate from unknown subtypes, prompting the utilization of mixture Gaussian graphical model for simultaneous subclassification and gene network inference. However, this method overlooks the heterogeneity of network relationships across subtypes and does not sufficiently emphasize shared relationships.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!