医学物理学
人工智能
计算机科学
医学
医学教育
心理学
作者
Ronald Chow,Shaakir Hasan,Ajay Zheng,Chenxi Gao,Gilmer Valdés,Francis T. S. Yu,Arpit M. Chhabra,Srinivas Raman,J. Isabelle Choi,Haibo Lin,Charles B. Simone
标识
DOI:10.1016/j.jacr.2024.07.011
摘要
The aim of this study is to assess the accuracy of ChatGPT in response to oncology exam questions in the setting of one-shot learning. Consecutive national radiation oncology in-service multiple-choice examinations were collected and inputted into ChatGPT 4o and ChatGPT 3.5 to determine ChatGPT's answers. ChatGPT's answers were then compared to the answer keys to determine whether ChatGPT correctly or incorrectly answered each question, and to determine if improvements in responses were seen with the newer ChatGPT version. A total of 600 consecutive questions were inputted into ChatGPT. ChatGPT 4o answered 72.2% questions correctly, whereas 3.5 answered 53.8% questions correctly. There was a significant difference in performance by question category (p < 0.01). ChatGPT performed poorer with respect to knowledge of landmark studies and treatment recommendations/planning. ChatGPT is a promising technology, with the latest version showing marked improvement. While it still has limitations, with further evolution, it may be considered a reliable resource for medical training and decision making in the oncology space.
科研通智能强力驱动
Strongly Powered by AbleSci AI