Can Patients Differentiate Responses by Orthopaedic Surgeons and ChatGPT Regarding Total Hip Arthroplasty?.
No Thumbnail Available
Authors
Prasad A.K.
Lim P.L.
Blackburn A.Z.
Cote M.P.
Ensor D.
Succi M.D.
Alpaugh K.
Melnic C.M.
Sepucha K.R.
Bedair, H. S.
Check for full-text access
Issue Date
2025
Type
Article
Language
Keywords
Alternative Title
Abstract
INTRODUCTION: The digital age offers vast health information, boosting patient health literacy but also fostering challenges like misinformation. Large language models-driven chatbots such as chat generative pretrained transformer 3 (ChatGPT 3) blur human artificial intelligence (AI) content creation boundaries, potentially affecting healthcare literacy. This study aimed to assess perceived message credibility by patients when comparing healthcare descriptions by ChatGPT 3 or orthopaedic attendings. METHOD(S): A cross-sectional survey of 160 patients assessed hip arthroplasty topics at increasing complexities: (1) "What is hip arthritis?"; (2) "What is a total hip replacement?"; (3) "What are the risks of undergoing a hip replacement?"; and (4) "What is the treatment of a periprosthetic joint infection of the hip?." Patients received one question randomly and compared four answers: one from ChatGPT and three from orthopaedic surgeons. Patients rated each answer (1 to 7) on credibility (accuracy, authenticity, believability) and indicated their most trusted answer. Multilevel modeling accounted for varying intercepts, providing estimated scores for each source. RESULT(S): Across all four questions, ChatGPT matched or exceeded the performance of orthopaedic surgeons across the domains of accuracy, authenticity, and believability. In three questions, ChatGPT surpassed at least one surgeon in one domain and, in two questions, outperformed at least two surgeons in two or more domains. Multilevel modeling revealed that ChatGPT scored the highest across all three credibility domains. Notably, ChatGPT's answer was the most trusted response by patients in two out of the four questions (Question #1 and #3). CONCLUSION(S): The study suggests that patients perceive ChatGPT as highly credible, particularly in terms of accuracy, authenticity, and believability compared with orthopaedic surgeons. These findings underscore the potential of AI to improve healthcare literacy and patient decision making. Further research is warranted to explore the nuances of patient trust and preferences in AI-generated healthcare content. LEVEL OF EVIDENCE: Level II, prospective, comparative study. Copyright © 2025 by the American Academy of Orthopaedic Surgeons.
Description
Citation
Publisher
License
Journal
The Journal of the American Academy of Orthopaedic Surgeons
Volume
Issue
Transcatheter Cardiovascular Therapeutics Abstracts.
