Background: Artificial intelligence (AI) systems hold great promise in improving medical care and health problems.
Aim: We aimed to evaluate the answers by asking the most frequently asked questions to ChatGPT for the prediction and treatment of fever, which is a major problem in children.
Methods: The 50 questions most frequently asked about fever in children were determined, and we asked them to ChatGPT. We evaluated the responses using the quality and readability scales.
Results: While ChatGPT demonstrated good quality in its responses, the readability scale and the Patient Education Material Evaluation Tool (PEMAT) scale used with materials appearing on the site were also found to be successful. Among the scales in which we evaluated ChatGPT responses, a weak positive relationship was found between Gunning Fog (GFOG) and Simple Measure of Gobbledygook (SMOG) scores (r = 0.379) and a significant and positive relationship was found between FGL and SMOG scores (r = 0.899).
Conclusion: This study sheds light on the quality and readability of information regarding the presentation of AI tools, such as ChatGPT, regarding fever, a common complaint in children. We determined that the answers to the most frequently asked questions about fire were high-quality, reliable, easy to read and understandable.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11701195 | PMC |
http://dx.doi.org/10.1111/jpc.16710 | DOI Listing |
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!