AI Article Synopsis

  • ChatGPT, developed by OpenAI, is a natural language processing model designed to improve human verbal communication, and this study tested its effectiveness in answering orthopedics-related multiple-choice questions (MCQs) and its role as a learning aid.
  • The study involved 129 undergraduate medical students who were split into two groups: one used ChatGPT as a study tool while the control group did not, with their understanding of orthopedics assessed after two weeks.
  • Results showed that ChatGPT had a 70.60% accuracy rate in answering 1051 MCQs, and those using ChatGPT scored higher on orthopedics tests compared to the control group, indicating its potential as a valuable learning resource.

Article Abstract

Background: ChatGPT is a natural language processing model developed by OpenAI, which can be iteratively updated and optimized to accommodate the changing and complex requirements of human verbal communication.

Objective: The study aimed to evaluate ChatGPT's accuracy in answering orthopedics-related multiple-choice questions (MCQs) and assess its short-term effects as a learning aid through a randomized controlled trial. In addition, long-term effects on student performance in other subjects were measured using final examination results.

Methods: We first evaluated ChatGPT's accuracy in answering MCQs pertaining to orthopedics across various question formats. Then, 129 undergraduate medical students participated in a randomized controlled study in which the ChatGPT group used ChatGPT as a learning tool, while the control group was prohibited from using artificial intelligence software to support learning. Following a 2-week intervention, the 2 groups' understanding of orthopedics was assessed by an orthopedics test, and variations in the 2 groups' performance in other disciplines were noted through a follow-up at the end of the semester.

Results: ChatGPT-4.0 answered 1051 orthopedics-related MCQs with a 70.60% (742/1051) accuracy rate, including 71.8% (237/330) accuracy for A1 MCQs, 73.7% (330/448) accuracy for A2 MCQs, 70.2% (92/131) accuracy for A3/4 MCQs, and 58.5% (83/142) accuracy for case analysis MCQs. As of April 7, 2023, a total of 129 individuals participated in the experiment. However, 19 individuals withdrew from the experiment at various phases; thus, as of July 1, 2023, a total of 110 individuals accomplished the trial and completed all follow-up work. After we intervened in the learning style of the students in the short term, the ChatGPT group answered more questions correctly than the control group (ChatGPT group: mean 141.20, SD 26.68; control group: mean 130.80, SD 25.56; P=.04) in the orthopedics test, particularly on A1 (ChatGPT group: mean 46.57, SD 8.52; control group: mean 42.18, SD 9.43; P=.01), A2 (ChatGPT group: mean 60.59, SD 10.58; control group: mean 56.66, SD 9.91; P=.047), and A3/4 MCQs (ChatGPT group: mean 19.57, SD 5.48; control group: mean 16.46, SD 4.58; P=.002). At the end of the semester, we found that the ChatGPT group performed better on final examinations in surgery (ChatGPT group: mean 76.54, SD 9.79; control group: mean 72.54, SD 8.11; P=.02) and obstetrics and gynecology (ChatGPT group: mean 75.98, SD 8.94; control group: mean 72.54, SD 8.66; P=.04) than the control group.

Conclusions: ChatGPT answers orthopedics-related MCQs accurately, and students using it excel in both short-term and long-term assessments. Our findings strongly support ChatGPT's integration into medical education, enhancing contemporary instructional methods.

Trial Registration: Chinese Clinical Trial Registry Chictr2300071774; https://www.chictr.org.cn/hvshowproject.html ?id=225740&v=1.0.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11372336PMC
http://dx.doi.org/10.2196/57037DOI Listing

Publication Analysis

Top Keywords

chatgpt group
36
control group
32
group
17
randomized controlled
12
chatgpt
12
mcqs
9
control
9
controlled trial
8
chatgpt's accuracy
8
accuracy answering
8

Similar Publications

Current status and associated factors of digital literacy among academic nurse educators: a cross-sectional study.

BMC Med Educ

January 2025

School of Nursing, Xiangnan University, 889 Chenzhou Avenue, Suxian District, Chenzhou, 423000, Hunan, People's Republic of China.

Background: In the backdrop of the ongoing global digital revolution in education, the digital literacy of teachers stands out as a pivotal determinant within the educational milieu. This study aims to explore the current status and associated factors of digital literacy among academic nurse educators.

Methods: A cross-sectional design study utilizing an online questionnaire platform (Wenjuanxing) to collect data from August to October 2023.

View Article and Find Full Text PDF

Introduction: Mental disorders, such as anxiety and depression, significantly impacted global populations in 2019 and 2020, with COVID-19 causing a surge in prevalence. They affect 13.4% of the people worldwide, and 21% of Iranians have experienced them.

View Article and Find Full Text PDF

Aging remains the foremost risk factor for cardiovascular and cerebrovascular diseases, surpassing traditional factors in epidemiological significance. This review elucidates the cellular and molecular mechanisms underlying vascular aging, with an emphasis on sex differences that influence disease progression and clinical outcomes in older adults. We discuss the convergence of aging processes at the macro- and microvascular levels and their contributions to the pathogenesis of vascular diseases.

View Article and Find Full Text PDF

Can ChatGPT be guide in pediatric dentistry?

BMC Oral Health

January 2025

Department of Pediatric Dentistry, Faculty of Dentistry, Tokat Gaziosmanpaşa University, Tokat, Türkiye.

Background: The use of ChatGPT in the field of health has recently gained popularity. In the field of dentistry, ChatGPT can provide services in areas such as, dental education and patient education. The aim of this study was to evaluate the quality, readability and originality of pediatric patient/parent information and academic content produced by ChatGPT in the field of pediatric dentistry.

View Article and Find Full Text PDF

Objective: With the development of ChatGPT, the number of studies within the nursing field has increased. The sophisticated language capabilities of ChatGPT, coupled with its exceptional precision, offer significant support within the nursing field, which includes clinical nursing, nursing education, and the clinical decision-making process. Preliminary findings suggest positive outcomes, underscoring its potential as a valuable resource for enhancing clinical care.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!