背景(考古学)
人工智能
医学
计算机科学
自然语言处理
机器学习
生物
古生物学
作者
Ruibin Feng,Kelly Brennan,Zahra Azizi,Jatin Goyal,Brototo Deb,Hui Ju Chang,Prasanth Ganesan,Paul Clopton,Maxime Pedron,Samuel Ruipérez-Campillo,Yaanik Desai,Hugo De Larochellière,Tina Baykaner,Marco Pérez,Rodrigo Bernardi Miguel,Albert J. Rogers,Sanjiv M. Narayan
出处
期刊:Circulation-arrhythmia and Electrophysiology
[Ovid Technologies (Wolters Kluwer)]
日期:2024-12-16
标识
DOI:10.1161/circep.124.013023
摘要
BACKGROUND: Large language models (LLMs), such as ChatGPT, excel at interpreting unstructured data from public sources, yet are limited when responding to queries on private repositories, such as electronic health records (EHRs). We hypothesized that prompt engineering could enhance the accuracy of LLMs for interpreting EHR data without requiring domain knowledge, thus expanding their utility for patients and personalized diagnostics. METHODS: We designed and systematically tested prompt engineering techniques to improve the ability of LLMs to interpret EHRs for nuanced diagnostic questions, referenced to a panel of medical experts. In 490 full-text EHR notes from 125 patients with prior life-threatening heart rhythm disorders, we asked GPT-4-turbo to identify recurrent arrhythmias distinct from prior events and tested 220 563 queries. To provide context, results were compared with rule-based natural language processing and BERT-based language models. Experiments were repeated for 2 additional LLMs. RESULTS: In an independent hold-out set of 389 notes, GPT-4-turbo had a balanced accuracy of 64.3%±4.7% out-of-the-box at baseline. This increased when asking GPT-4-turbo to provide a rationale for its answers, requiring a structured data output, and providing in-context exemplars, rose to a balanced accuracy of 91.4%±3.8% ( P <0.05). This surpassed the traditional logic-based natural language processing and BERT-based models ( P <0.05). Results were consistent for GPT-3.5-turbo and Jurassic-2 LLMs. CONCLUSIONS: The use of prompt engineering strategies enables LLMs to identify clinical end points from EHRs with an accuracy that surpassed natural language processing and approximated experts, yet without the need for expert knowledge. These approaches could be applied to LLM queries for other domains, to facilitate automated analysis of nuanced data sets with high accuracy by nonexperts.
科研通智能强力驱动
Strongly Powered by AbleSci AI