亲爱的研友该休息了!由于当前在线用户较少,发布求助请尽量完整地填写文献信息,科研通机器人24小时在线,伴您度过漫漫科研夜!身体可是革命的本钱,早点休息,好梦!

ChatGPT4 Outperforms Endoscopists for Determination of Postcolonoscopy Rescreening and Surveillance Recommendations

医学 结肠镜检查 一致性 麦克内马尔试验 指南 临床实习 置信区间 结直肠癌筛查 普通外科 家庭医学 医学物理学 胃肠病学 内科学 病理 结直肠癌 癌症 统计 数学
作者
Patrick Chang,Maziar M. Amini,Rio O. Davis,Denis Nguyen,Jennifer L. Dodge,Helen Lee,Sarah Sheibani,Jennifer Phan,James Buxbaum,Ara Sahakian
出处
期刊:Clinical Gastroenterology and Hepatology [Elsevier]
卷期号:22 (9): 1917-1925.e17 被引量:21
标识
DOI:10.1016/j.cgh.2024.04.022
摘要

Background Large language models (LLM) including ChatGPT4 improve access to artificial intelligence, but their impact on the clinical practice of gastroenterology is undefined. In this study, we aim to compare the accuracy, concordance and reliability of ChatGPT4 colonoscopy recommendations for colorectal cancer re-screening and surveillance to contemporary guidelines and real-world gastroenterology practice. Methods History of present illness, colonoscopy data and pathology reports from patients undergoing procedures at two large academic centers were entered into ChatGPT4 and it was queried for next recommended colonoscopy follow-up interval. Using McNemar's test and inter-rater reliability, we compared the recommendations made by ChatGPT4 with the actual surveillance interval provided in the endoscopist's procedure report (gastroenterology practice) and the appropriate USMSTF guidance. The latter was generated for each case by an expert panel using the clinical information and guideline documents as reference. Results Text input of de-identified data into ChatGPT4 from 505 consecutive patients undergoing colonoscopy between January 1st and April 30th, 2023 elicited a successful follow-up recommendation in 99.2% of the queries. ChatGPT4 recommendations were in closer agreement with the USMSTF Panel (85.7%) than gastroenterology practice recommendations with the USMSTF Panel (75.4%) (P<.001). Of the 14.3% discordant recommendations between ChatGPT4 and USMSTF Panel, recommendations were for later screening in 26 (5.1%) and earlier screening in 44 (8.7%) cases. The inter-rater reliability was good for ChatGPT4 vs. USMSTF Panel (Fleiss κ: 0.786, CI95%: 0.734-0.838, P<.001). Conclusions Initial real-world results suggest that ChatGPT4 can accurately define routine colonoscopy screening intervals based on verbatim input of clinical data. LLM have potential for clinical applications, but further training is needed for broad use. Large language models (LLM) including ChatGPT4 improve access to artificial intelligence, but their impact on the clinical practice of gastroenterology is undefined. In this study, we aim to compare the accuracy, concordance and reliability of ChatGPT4 colonoscopy recommendations for colorectal cancer re-screening and surveillance to contemporary guidelines and real-world gastroenterology practice. History of present illness, colonoscopy data and pathology reports from patients undergoing procedures at two large academic centers were entered into ChatGPT4 and it was queried for next recommended colonoscopy follow-up interval. Using McNemar's test and inter-rater reliability, we compared the recommendations made by ChatGPT4 with the actual surveillance interval provided in the endoscopist's procedure report (gastroenterology practice) and the appropriate USMSTF guidance. The latter was generated for each case by an expert panel using the clinical information and guideline documents as reference. Text input of de-identified data into ChatGPT4 from 505 consecutive patients undergoing colonoscopy between January 1st and April 30th, 2023 elicited a successful follow-up recommendation in 99.2% of the queries. ChatGPT4 recommendations were in closer agreement with the USMSTF Panel (85.7%) than gastroenterology practice recommendations with the USMSTF Panel (75.4%) (P<.001). Of the 14.3% discordant recommendations between ChatGPT4 and USMSTF Panel, recommendations were for later screening in 26 (5.1%) and earlier screening in 44 (8.7%) cases. The inter-rater reliability was good for ChatGPT4 vs. USMSTF Panel (Fleiss κ: 0.786, CI95%: 0.734-0.838, P<.001). Initial real-world results suggest that ChatGPT4 can accurately define routine colonoscopy screening intervals based on verbatim input of clinical data. LLM have potential for clinical applications, but further training is needed for broad use.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
2秒前
发nature发布了新的文献求助10
5秒前
10秒前
Sneijder10应助提米橘采纳,获得10
29秒前
31秒前
西升东落完成签到,获得积分10
35秒前
Sneijder10发布了新的文献求助50
35秒前
西升东落发布了新的文献求助20
39秒前
研友_VZG7GZ应助风花雪月采纳,获得10
40秒前
小橘子吃傻子完成签到,获得积分10
42秒前
Sneijder10应助提米橘采纳,获得10
43秒前
Sneijder10应助提米橘采纳,获得10
54秒前
Alisha完成签到,获得积分10
55秒前
GPTea应助西升东落采纳,获得20
55秒前
打打应助发nature采纳,获得10
1分钟前
1分钟前
YIZEXIN发布了新的文献求助10
1分钟前
鹭江发布了新的文献求助30
1分钟前
1分钟前
1分钟前
风花雪月发布了新的文献求助10
1分钟前
Ciwei发布了新的文献求助10
1分钟前
1分钟前
发nature发布了新的文献求助10
1分钟前
1分钟前
顾矜应助Ciwei采纳,获得10
1分钟前
sanages发布了新的文献求助10
1分钟前
Orange应助发nature采纳,获得10
1分钟前
充电宝应助尊敬的半梅采纳,获得10
1分钟前
慕青应助YIZEXIN采纳,获得10
2分钟前
2分钟前
发nature发布了新的文献求助10
2分钟前
2分钟前
桐桐应助科研通管家采纳,获得10
2分钟前
FashionBoy应助科研通管家采纳,获得10
2分钟前
2分钟前
可千万不要躺平呀完成签到,获得积分0
2分钟前
YIZEXIN发布了新的文献求助10
2分钟前
2分钟前
2分钟前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Handbook of pharmaceutical excipients, Ninth edition 5000
Aerospace Standards Index - 2026 ASIN2026 3000
Relation between chemical structure and local anesthetic action: tertiary alkylamine derivatives of diphenylhydantoin 1000
Signals, Systems, and Signal Processing 610
Discrete-Time Signals and Systems 610
Principles of town planning : translating concepts to applications 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 工程类 纳米技术 有机化学 物理 生物化学 化学工程 计算机科学 复合材料 内科学 催化作用 光电子学 物理化学 电极 冶金 遗传学 细胞生物学
热门帖子
关注 科研通微信公众号,转发送积分 6066173
求助须知:如何正确求助?哪些是违规求助? 7898430
关于积分的说明 16322665
捐赠科研通 5208268
什么是DOI,文献DOI怎么找? 2786257
邀请新用户注册赠送积分活动 1768997
关于科研通互助平台的介绍 1647799