Evaluation of Chat Generative Pre-trained Transformer’s responses to frequently asked questions about psoriatic arthritis: A study on quality and readability
Keywords:
ChatGPT, Artificial intelligence, Psoriatic arthritis, Quality information, ReadabilityAbstract
Aim: The growing use of artificial intelligence (AI) in healthcare, especially through technologies such as Chat Generative Pre-trained Transformer (ChatGPT), has led to concerns regarding the quality and readability of AI-generated health data. This study aimed to evaluate ChatGPT’s responses to frequently asked questions about psoriatic arthritis (PsA).
Materials and Methods: The quality of ChatGPT-generated responses was evaluated using the Ensuring Quality Information for Patients (EQIP) tool. Readability was assessed using the Flesch–Kincaid Reading Ease (FKRE) and Flesch–Kincaid Grade Level (FKGL) indices. The Kruskal–Wallis H test was used to compare subgroups, and Bonferroni correction was done for multiple comparisons.
Results: Significant differences were observed in EQIP scores across question subgroups, with treatment-related questions scoring lower than symptom-related questions. The FKRE and FKGL scores indicated that the information provided by ChatGPT could be challenging for patients with lower literacy levels.
Conclusion: Although ChatGPT provided relatively accurate information on PsA, its readability and ability to communicate complex medical information might be improved. These findings suggest the necessity for continual refinement of AI tools to address the diverse needs of patients.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Annals of Medical Research

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
CC Attribution-NonCommercial-NoDerivatives 4.0