RSS-Feed abonnieren

DOI: 10.1055/a-2565-9155
Automating Responses to Patient Portal Messages Using Generative AI
Funding None.

Abstract
Background
Patient portals bridge patient and provider communications but exacerbate physician and nursing burnout. Large language models (LLMs) can generate message responses that are viewed favorably by health care professionals/providers (HCPs); however, these studies have not included diverse message types or new prompt-engineering strategies.
Objectives
Our goal is to investigate and compare the quality and precision of GPT-generated message responses versus real doctor responses across the spectrum of message types within a patient portal.
Methods
We used prompt engineering techniques to craft synthetic provider responses tailored to adult primary care patients. We enrolled a sample of primary care providers in a cross-sectional study to compare authentic with synthetic patient portal message responses generated by GPT-3.5-turbo, July 2023 version (GPT). The survey assessed each response's empathy, relevance, medical accuracy, and readability on a scale from 0 to 5. Respondents were asked to identify responses that were GPT-generated versus provider-generated. Mean scores for all metrics were computed for subsequent analysis.
Results
A total of 49 HCPs participated in the survey (59% completion rate), comprising 16 physicians and 32 advanced practice providers (APPs). In comparison to responses generated by real doctors, GPT-generated responses scored statistically significantly higher than doctors in two of the four parameters: empathy (p < 0.05) and readability (p < 0.05). However, no statistically significant difference was observed for relevance and accuracy (p > 0.05). Although readability scores were significantly different, the absolute difference was small, and the clinical significance of this finding remains uncertain.
Conclusion
Our findings affirm the potential of GPT-generated message responses to achieve comparable levels of empathy, relevance, and readability to those found in typical responses crafted by HCPs. Additional studies should be done within provider workflows and with careful evaluation of patient attitudes and concerns related to the ethics as well as the quality of generated responses in all settings.
Keywords
patient web portal - physician–patient interaction - artificial intelligence - communication - health - medical informatics - large language modelsProtection of Human Subjects
The University of Pennsylvania Human Research Protection Program, under study No. 854147, granted approval for this research project. Participant consent was not deemed necessary as the study involved secondary data analysis of patient portal messages sourced through a meticulously crafted pipeline. Furthermore, the protocol for this research, also approved under study No. 854147, granted approval for retrieving the initial set of patient portal messages from a repository at Vanderbilt University Medical Center (VUMC), which were later used to create synthetic patient portal messages used in the study. The utilization of patient portal messages from VUMC was conducted in compliance with ethical guidelines. This study did not require patient consent for using the patient portal messages retrieved from VUMC, as the data used in this study underwent a rigorous de-identification process, rendering it impossible to trace any information back to individual patients. Thus, our research respects and upholds the principles of confidentiality and anonymity, ensuring the protection of participants' privacy rights in accordance with established ethical standards.
Publikationsverlauf
Eingereicht: 12. August 2024
Angenommen: 11. Februar 2025
Accepted Manuscript online:
25. März 2025
Artikel online veröffentlicht:
30. Juli 2025
© 2025. The Author(s). This is an open access article published by Thieme under the terms of the Creative Commons Attribution License, permitting unrestricted use, distribution, and reproduction so long as the original work is properly cited. (https://creativecommons.org/licenses/by/4.0/)
Georg Thieme Verlag KG
Oswald-Hesse-Straße 50, 70469 Stuttgart, Germany
-
References
- 1 Carini E, Villani L, Pezzullo AM. et al. The impact of digital patient portals on health outcomes, system efficiency, and patient attitudes: Updated systematic literature review. J Med Internet Res 2021; 23 (09) e26189
- 2 Tai-Seale M, Baxter S, Millen M. et al. Association of physician burnout with perceived EHR work stress and potentially actionable factors. J Am Med Inform Assoc 2023; 30 (10) 1665-1672
- 3 Johnson KB, Neuss MJ, Detmer DE. Electronic health records and clinician burnout: A story of three eras. J Am Med Inform Assoc 2021; 28 (05) 967-973
- 4 Johnson KB, Ibrahim SA, Rosenbloom ST. Ensuring equitable access to patient portals-closing the “techquity” gap. JAMA Health Forum 2023; 4 (11) e233406
- 5 Kruse CS, Mileski M, Dray G, Johnson Z, Shaw C, Shirodkar H. Physician burnout and the electronic health record leading up to and during the first year of COVID-19: Systematic review. J Med Internet Res 2022; 24 (03) e36200
- 6 Liu S, Wright AP, Patterson BL. et al. Using AI-generated suggestions from ChatGPT to optimize clinical decision support. J Am Med Inform Assoc 2023; 30 (07) 1237-1245
- 7 Rajjoub R, Arroyave JS, Zaidat B. et al. ChatGPT and its role in the decision-making for the diagnosis and treatment of lumbar spinal stenosis: A comparative analysis and narrative review. Global Spine J 2024; 14 (03) 998-1017
- 8 Kao HJ, Chien TW, Wang WC, Chou W, Chow JC. Assessing ChatGPT's capacity for clinical decision support in pediatrics: A comparative study with pediatricians using KIDMAP of Rasch analysis. Medicine (Baltimore) 2023; 102 (25) e34068
- 9 Liu S, McCoy AB, Wright AP. et al. Leveraging large language models for generating responses to patient messages. medRxiv 2023 (e-pub ahead of print)
- 10 Liu J, Wang C, Liu S. Utility of ChatGPT in clinical practice. J Med Internet Res 2023; 25: e48568
- 11 Kurniawan MH, Handiyani H, Nuraini T, Hariyati RTS, Sutrisno S. A systematic review of artificial intelligence-powered (AI-powered) chatbot intervention for managing chronic illness. Ann Med 2024; 56 (01) 2302980
- 12 Mohan RMR, Joy M, Natt D. et al. Digital Therapeutics and Chatbots: Assessing the efficacy of Chat GPT and Google BARD in IBS treatment plans. Am J Gastroenterol 2024; 118: S25
- 13 Chheng C, Wilson D. Abnormal gait detection using wearable Hall-Effect sensors. Sensors (Basel) 2021; 21 (04) 1206
- 14 Ram Mohan RM, Joy M, Natt D. et al. The Future Of Digestive Health: Personalized treatments through wearable AI technologies. Endoscopy 2024; 56: S316
- 15 Soley N, Speed TJ, Xie A, Taylor CO. Predicting postoperative pain and opioid use with machine learning applied to longitudinal electronic health record and wearable data. Appl Clin Inform 2024; 15 (03) 569-582
- 16 Tai-Seale M, Baxter SL, Vaida F. et al. AI-generated draft replies integrated into health records and physicians' electronic communication. JAMA Netw Open 2024; 7 (04) e246565
- 17 Sulieman L, Robinson JR, Jackson GP. Automating the classification of complexity of medical decision-making in patient-provider messaging in a patient portal. J Surg Res 2020; 255: 224-232
- 18 Achiam J, Adler S, Agarwal S. et al.; Open AI. GPT-4 technical report. Published online 2023. arXiv:2303:08774
- 19 Narang S, Raffel C, Lee K, Roberts A, Fiedel N, Malkan K. WT5?! Training Text-to-text models to explain their predictions. Published online April 29, 2020. arXiv:2004:14546. Accessed August 5, 2024 at: http://arxiv.org/abs/2004.14546
- 20 Wiegreffe S, Hessel J, Swayamdipta S, Riedl M, Choi Y. Reframing human-AI collaboration for generating free-text explanations. Published online May 4, 2022. arXiv:2112:08674. Accessed August 5, 2024 at: http://arxiv.org/abs/2112.08674
- 21 Budko A. Observer Project Github site. Published online November 18, 2024. Accessed April 4, 2025 at: https://github.com/alex-budko/OBSERVER-Project
- 22 Ayers JW, Poliak A, Dredze M. et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med 2023; 183 (06) 589-596
- 23 Garcia P, Ma SP, Shah S. et al. Artificial intelligence-generated draft replies to patient inbox messages. JAMA Netw Open 2024; 7 (03) e243201
- 24 Riedel M, Kaefinger K, Stuehrenberg A. et al. ChatGPT's performance in German OB/GYN exams - paving the way for AI-enhanced medical education and clinical practice. Front Med (Lausanne) 2023; 10: 1296615
- 25 Reynolds K, Tejasvi T. Potential use of ChatGPT in responding to patient questions and creating patient resources. JMIR Dermatol 2024; 7: e48451