Recent advancements in large language models (LLMs) such as ChatGPT and LLaMA have hinted at their potential to revolutionize medical applications, yet their application in clinical settings often reveals limitations due to a lack of specialized training on medical-specific data. In response to this challenge, this study introduces Me-LLaMA, a novel medical LLM family that includes foundation models - Me-LLaMA 13/70B, along with their chat-enhanced versions - Me-LLaMA 13/70B-chat, developed through continual pre-training and instruction tuning of LLaMA2 using large medical datasets. Our methodology leverages a comprehensive domain-specific data suite, including a large-scale, continual pre-training dataset with 129B tokens, an instruction tuning dataset with 214k samples, and a new medical evaluation benchmark (MIBE) across six critical medical tasks with 12 datasets. Our extensive evaluation using the MIBE shows that Me-LLaMA models achieve overall better performance than existing open-source medical LLMs in zero-shot, few-shot and supervised learning abilities. With task-specific instruction tuning, Me-LLaMA models outperform ChatGPT on 7 out of 8 datasets and GPT-4 on 5 out of 8 datasets. In addition, we investigated the catastrophic forgetting problem, and our results show that Me-LLaMA models outperform other open-source medical LLMs in mitigating this issue. Me-LLaMA is one of the largest open-source medical foundation LLMs that use both biomedical and clinical data. It exhibits superior performance across both general and medical tasks compared to other open-source medical LLMs, rendering it an attractive choice for medical AI applications. We release our models, datasets, and evaluation scripts at: https://github.com/BIDS-Xu-Lab/Me-LLaMA.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11142305PMC
http://dx.doi.org/10.21203/rs.3.rs-4240043/v1DOI Listing

Publication Analysis

Top Keywords

open-source medical
16
medical
12
medical applications
12
instruction tuning
12
me-llama models
12
medical llms
12
me-llama
8
large language
8
language models
8
continual pre-training
8

Similar Publications

Objective: Medical laboratory data together with prescribing and hospitalisation records are three of the most used electronic health records (EHRs) for data-driven health research. In Scotland, hospitalisation, prescribing and the death register data are available nationally whereas laboratory data is captured, stored and reported from local health board systems with significant heterogeneity. For researchers or other users of this regionally curated data, working on laboratory datasets across regional cohorts requires effort and time.

View Article and Find Full Text PDF

Spatially resolved omics (SRO) technologies enable the identification of cell types while preserving their organization within tissues. Application of such technologies offers the opportunity to delineate cell-type spatial relationships, particularly across different length scales, and enhance our understanding of tissue organization and function. To quantify such multi-scale cell-type spatial relationships, we present CRAWDAD, Cell-type Relationship Analysis Workflow Done Across Distances, as an open-source R package.

View Article and Find Full Text PDF
Article Synopsis
  • Acute kidney injury (AKI) affects a significant number of critically ill patients, with the lack of standardized tools for implementing KDIGO criteria creating challenges for researchers.
  • The pyAKI pipeline was developed to address these issues, using the MIMIC-IV database to establish a standardized model for consistent AKI diagnosis.
  • Validation tests showed that pyAKI performs better than human annotations, achieving perfect accuracy and offering a valuable resource for clinicians and data scientists in AKI research.
View Article and Find Full Text PDF

Microsatellite instability (MSI) is a critical phenotype of cancer genomes and an FDA-recognized biomarker that can guide treatment with immune checkpoint inhibitors. Previous work has demonstrated that next-generation sequencing data can be used to identify samples with MSI-high phenotype. However, low tumor purity, as frequently observed in routine clinical samples, poses a challenge to the sensitivity of existing algorithms.

View Article and Find Full Text PDF

Cost-effectiveness models of non-small cell lung cancer: A systematic literature review.

J Manag Care Spec Pharm

January 2025

Medical Oncology Department 1, Fondazione IRCCS Istituto Nazionale Tumori, Milan, Italy (Prelaj).

Background: Non-small cell lung cancer (NSCLC) presents a formidable global health challenge owing to significant morbidity, high mortality rates, and substantial economic burden. Recent advances in targeted therapies and immunotherapies have transformed NSCLC treatment, but efficacy varies across patients. Tailoring treatment to patients can improve outcomes and potentially improve cost-effectiveness (ie, value for money) as well.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!