Leveraging large language models to construct feedback from medical multiple-choice Questions.

Sci Rep

Data-Intensive Systems and Visualization Group (dAI.SY), Fakultät für Informatik und Automatisierung, Technische Universität Ilmenau, Ehrenbergstraße 29, 98693, Ilmenau, Thuringia, Germany.

Published: November 2024

Exams like the formative Progress Test Medizin can enhance their effectiveness by offering feedback beyond numerical scores. Content-based feedback, which encompasses relevant information from exam questions, can be valuable for students by offering them insight into their performance on the current exam, as well as serving as study aids and tools for revision. Our goal was to utilize Large Language Models (LLMs) in preparing content-based feedback for the Progress Test Medizin and evaluate their effectiveness in this task. We utilize two popular LLMs and conduct a comparative assessment by performing textual similarity on the generated outputs. Furthermore, we study via a survey how medical practitioners and medical educators assess the capabilities of LLMs and perceive the usage of LLMs for the task of generating content-based feedback for PTM exams. Our findings show that both examined LLMs performed similarly. Both have their own advantages and disadvantages. Our survey results indicate that one LLM produces slightly better outputs; however, this comes at a cost since it is a paid service, while the other is free to use. Overall, medical practitioners and educators who participated in the survey find the generated feedback relevant and useful, and they are open to using LLMs for such tasks in the future. We conclude that while the content-based feedback generated by the LLM may not be perfect, it nevertheless can be considered a valuable addition to the numerical feedback currently provided.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11561272PMC
http://dx.doi.org/10.1038/s41598-024-79245-xDOI Listing

Publication Analysis

Top Keywords

content-based feedback
16
large language
8
language models
8
feedback
8
progress test
8
test medizin
8
medical practitioners
8
llms
6
leveraging large
4
models construct
4

Similar Publications

Background: Following the US Supreme Court decision overturning Roe v. Wade, there is evidence of limitations in access to safe abortion care. Artificially intelligent (AI)-enabled conversational chatbots are becoming an appealing option to support access to care, but generative AI systems can misinform and hallucinate and risk reinforcing problematic bias and stigma related to sexual and reproductive healthcare.

View Article and Find Full Text PDF

Leveraging large language models to construct feedback from medical multiple-choice Questions.

Sci Rep

November 2024

Data-Intensive Systems and Visualization Group (dAI.SY), Fakultät für Informatik und Automatisierung, Technische Universität Ilmenau, Ehrenbergstraße 29, 98693, Ilmenau, Thuringia, Germany.

Exams like the formative Progress Test Medizin can enhance their effectiveness by offering feedback beyond numerical scores. Content-based feedback, which encompasses relevant information from exam questions, can be valuable for students by offering them insight into their performance on the current exam, as well as serving as study aids and tools for revision. Our goal was to utilize Large Language Models (LLMs) in preparing content-based feedback for the Progress Test Medizin and evaluate their effectiveness in this task.

View Article and Find Full Text PDF

Adaptation of mental health first aid guidelines for eating disorders for Iran.

BMC Psychiatry

September 2024

School of Population and Global Health, Centre for Mental Health, The University of Melbourne, MelbourneMelbourne, Australia.

Background: This study aimed to adapt mental health first aid guidelines to support individuals with or at risk of developing eating disorders in Iran. This adaptation seeks to enhance the support available for the Iranian population dealing with these disorders.

Methods: We employed the Delphi expert consensus method, utilizing two panels: health professionals (n = 37 in the first round; n = 29 in the second) and individuals with lived experience (n = 20 in the first round; n = 18 in the second).

View Article and Find Full Text PDF
Article Synopsis
  • The study evaluates a virtual reality game called Golden Breath, designed to reduce pain and fear in children undergoing needle procedures by using biofeedback-based techniques.
  • Development involved various phases, including feature assessment, gamification, expert feedback, and usability testing with 11 children aged 4-12 years, all of whom reported high satisfaction.
  • Results indicated that Golden Breath is feasible, acceptable, and safe, effectively reducing children's pain and fear during needle-related procedures, suggesting it should be integrated into healthcare practices for pediatric patients.
View Article and Find Full Text PDF

Virtual Waiting Room: The New Narrative of Waiting in Oncology Care.

J Cancer Educ

September 2024

Department of Medical Psychology and Psychosomatics, Faculty of Medicine, Masaryk University, Brno, Czechia.

This conceptual study introduces the "virtual waiting room," an innovative, interactive, web-based platform designed to enhance the waiting experience in oncology by providing personalized, educational, and supportive content. Central to our study is the implementation of the circular entry model, which allows for non-linear navigation of health information, empowering patients to access content based on their immediate needs and interests. This approach respects the individual journeys of patients, acknowledging the diverse pathways through which they seek understanding and manage their health.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!