Evaluation of information provided to patients by ChatGPT about chronic diseases in Spanish language

Introduction Artificial intelligence has presented exponential growth in medicine. The ChatGPT language model has been highlighted as a possible source of patient information. This study evaluates the reliability and readability of ChatGPT-generated patient information on chronic diseases in Spanish...

Full description

Bibliographic Details
Main Authors: María Juliana Soto-Chávez, Marlon Mauricio Bustos, Daniel G. Fernández-Ávila, Oscar Mauricio Muñoz
Format: Article
Language:English
Published: SAGE Publishing 2024-01-01
Series:Digital Health
Online Access:https://doi.org/10.1177/20552076231224603
_version_ 1827390834816843776
author María Juliana Soto-Chávez
Marlon Mauricio Bustos
Daniel G. Fernández-Ávila
Oscar Mauricio Muñoz
author_facet María Juliana Soto-Chávez
Marlon Mauricio Bustos
Daniel G. Fernández-Ávila
Oscar Mauricio Muñoz
author_sort María Juliana Soto-Chávez
collection DOAJ
description Introduction Artificial intelligence has presented exponential growth in medicine. The ChatGPT language model has been highlighted as a possible source of patient information. This study evaluates the reliability and readability of ChatGPT-generated patient information on chronic diseases in Spanish. Methods Questions frequently asked by patients on the internet about diabetes mellitus, heart failure, rheumatoid arthritis (RA), chronic kidney disease (CKD), and systemic lupus erythematosus (SLE) were submitted to ChatGPT. Reliability was assessed by rating responses as (1) comprehensive, (2) correct but inadequate, (3) some correct and some incorrect, (4) completely incorrect, and divided between “good” (1 and 2) and “bad” (3 and 4). Readability was evaluated with the adapted Flesch and Szigriszt formulas. Results And 71.67% of the answers were “good,” with none qualified as “completely incorrect.” Better reliability was observed in questions on diabetes and RA versus heart failure (p = 0.02). In readability, responses were “moderately difficult” (54.73, interquartile range (IQR) 51.59–58.58), with better results for CKD (median 56.1, IQR 53.5–59.1) and RA (56.4, IQR 53.7–60.7), than for heart failure responses (median 50.6, IQR 46.3–53.8). Conclusion Our study suggests that the ChatGPT tool can be a reliable source of information in spanish for patients with chronic diseases with different reliability for some of them, however, it needs to improve the readability of its answers to be recommended as a useful tool for patients.
first_indexed 2024-03-08T16:58:35Z
format Article
id doaj.art-293e9ca2d2074c9a9c8bf469e88fd219
institution Directory Open Access Journal
issn 2055-2076
language English
last_indexed 2024-03-08T16:58:35Z
publishDate 2024-01-01
publisher SAGE Publishing
record_format Article
series Digital Health
spelling doaj.art-293e9ca2d2074c9a9c8bf469e88fd2192024-01-04T16:07:25ZengSAGE PublishingDigital Health2055-20762024-01-011010.1177/20552076231224603Evaluation of information provided to patients by ChatGPT about chronic diseases in Spanish languageMaría Juliana Soto-Chávez0Marlon Mauricio Bustos1 Daniel G. Fernández-Ávila2Oscar Mauricio Muñoz3 Department of Internal Medicine, , Bogotá, Colombia Department of Internal Medicine, , Bogotá, Colombia Rheumatology Unit, , Bogotá, Colombia Department of Internal Medicine, , Bogotá, ColombiaIntroduction Artificial intelligence has presented exponential growth in medicine. The ChatGPT language model has been highlighted as a possible source of patient information. This study evaluates the reliability and readability of ChatGPT-generated patient information on chronic diseases in Spanish. Methods Questions frequently asked by patients on the internet about diabetes mellitus, heart failure, rheumatoid arthritis (RA), chronic kidney disease (CKD), and systemic lupus erythematosus (SLE) were submitted to ChatGPT. Reliability was assessed by rating responses as (1) comprehensive, (2) correct but inadequate, (3) some correct and some incorrect, (4) completely incorrect, and divided between “good” (1 and 2) and “bad” (3 and 4). Readability was evaluated with the adapted Flesch and Szigriszt formulas. Results And 71.67% of the answers were “good,” with none qualified as “completely incorrect.” Better reliability was observed in questions on diabetes and RA versus heart failure (p = 0.02). In readability, responses were “moderately difficult” (54.73, interquartile range (IQR) 51.59–58.58), with better results for CKD (median 56.1, IQR 53.5–59.1) and RA (56.4, IQR 53.7–60.7), than for heart failure responses (median 50.6, IQR 46.3–53.8). Conclusion Our study suggests that the ChatGPT tool can be a reliable source of information in spanish for patients with chronic diseases with different reliability for some of them, however, it needs to improve the readability of its answers to be recommended as a useful tool for patients.https://doi.org/10.1177/20552076231224603
spellingShingle María Juliana Soto-Chávez
Marlon Mauricio Bustos
Daniel G. Fernández-Ávila
Oscar Mauricio Muñoz
Evaluation of information provided to patients by ChatGPT about chronic diseases in Spanish language
Digital Health
title Evaluation of information provided to patients by ChatGPT about chronic diseases in Spanish language
title_full Evaluation of information provided to patients by ChatGPT about chronic diseases in Spanish language
title_fullStr Evaluation of information provided to patients by ChatGPT about chronic diseases in Spanish language
title_full_unstemmed Evaluation of information provided to patients by ChatGPT about chronic diseases in Spanish language
title_short Evaluation of information provided to patients by ChatGPT about chronic diseases in Spanish language
title_sort evaluation of information provided to patients by chatgpt about chronic diseases in spanish language
url https://doi.org/10.1177/20552076231224603
work_keys_str_mv AT mariajulianasotochavez evaluationofinformationprovidedtopatientsbychatgptaboutchronicdiseasesinspanishlanguage
AT marlonmauriciobustos evaluationofinformationprovidedtopatientsbychatgptaboutchronicdiseasesinspanishlanguage
AT danielgfernandezavila evaluationofinformationprovidedtopatientsbychatgptaboutchronicdiseasesinspanishlanguage
AT oscarmauriciomunoz evaluationofinformationprovidedtopatientsbychatgptaboutchronicdiseasesinspanishlanguage