Article
作者: Roginiel, Aliya C ; Kalaw, Fritz Gerald P. ; Scott, Nathan L. ; Chen, Jimmy S ; Koretz, Zachary A ; Koretz, Zachary A. ; Reddy, Akshay J. ; Weinreb, Robert N. ; Baxter, Sally L. ; Robbins, Shira L ; Do, Jiun L. ; Lang, Paul Z ; Lang, Paul Z. ; Kalaw, Fritz Gerald P ; Al-Sharif, Eman ; Granet, David B ; Scott, Nathan L ; Chen, Jimmy S. ; Arnett, Justin J ; Camp, Andrew S. ; Bolo, Kyle A ; Bolo, Kyle A. ; Weinreb, Robert N ; Shoji, Marissa K ; Granet, David B. ; Robbins, Shira L. ; Rudell, Jolene C. ; Camp, Andrew S ; Arya, Malvika ; Do, Jiun L ; Rudell, Jolene C ; Roginiel, Aliya C. ; Shoji, Marissa K. ; Arnett, Justin J. ; Reddy, Akshay J ; Baxter, Sally L ; Eslani, Medi
ObjectiveLarge language models such as ChatGPT have demonstrated significant potential in question-answering within ophthalmology, but there is a paucity of literature evaluating its ability to generate clinical assessments and discussions. The objectives of this study were to (1) assess the accuracy of assessment and plans generated by ChatGPT and (2) evaluate ophthalmologists' abilities to distinguish between responses generated by clinicians versus ChatGPT.DesignCross-sectional mixed-methods study.SubjectsSixteen ophthalmologists from a single academic center, of which 10 were board-eligible and 6 were board-certified, were recruited to participate in this study.MethodsPrompt engineering was used to ensure ChatGPT output discussions in the style of the ophthalmologist author of the Medical College of Wisconsin Ophthalmic Case Studies. Cases where ChatGPT accurately identified the primary diagnoses were included and then paired. Masked human-generated and ChatGPT-generated discussions were sent to participating ophthalmologists to identify the author of the discussions. Response confidence was assessed using a 5-point Likert scale score, and subjective feedback was manually reviewed.Main Outcome MeasuresAccuracy of ophthalmologist identification of discussion author, as well as subjective perceptions of human-generated versus ChatGPT-generated discussions.ResultsOverall, ChatGPT correctly identified the primary diagnosis in 15 of 17 (88.2%) cases. Two cases were excluded from the paired comparison due to hallucinations or fabrications of nonuser-provided data. Ophthalmologists correctly identified the author in 77.9% ± 26.6% of the 13 included cases, with a mean Likert scale confidence rating of 3.6 ± 1.0. No significant differences in performance or confidence were found between board-certified and board-eligible ophthalmologists. Subjectively, ophthalmologists found that discussions written by ChatGPT tended to have more generic responses, irrelevant information, hallucinated more frequently, and had distinct syntactic patterns (all P < 0.01).ConclusionsLarge language models have the potential to synthesize clinical data and generate ophthalmic discussions. While these findings have exciting implications for artificial intelligence-assisted health care delivery, more rigorous real-world evaluation of these models is necessary before clinical deployment.Financial DisclosuresThe author(s) have no proprietary or commercial interest in any materials discussed in this article.