已入深夜,您辛苦了!由于当前在线用户较少,发布求助请尽量完整地填写文献信息,科研通机器人24小时在线,伴您度过漫漫科研夜!祝你早点完成任务,早点休息,好梦!

Bias in Artificial Intelligence

医学 机器学习 种族(生物学) 深度学习 肾移植 肾病科 肾移植 精密医学 人工神经网络 人工智能 数据科学 移植 内科学 病理 计算机科学 植物 生物
作者
Yoonyoung Park,Jianying Hu
出处
期刊:Clinical Journal of The American Society of Nephrology [American Society of Nephrology]
卷期号:18 (3): 394-396 被引量:1
标识
DOI:10.2215/cjn.0000000000000078
摘要

Recent advances in machine learning and artificial intelligence (AI) brought unprecedented promises across the fields of medicine, including nephrology. The clinical complexity and challenges in patient management highlight the potential benefit of data-driven, algorithmic approaches in nephrology.1 For example, neural networks and other deep learning methods have been applied, from analyzing kidney biopsy specimens to predicting kidney failures.2 Along with hopes and hype comes the increasing concern that data- and model-based decision making can in fact exacerbate bias and inequity in health care. Researchers have shown that a model-driven prediction of eGFR that has been used for decades could be racially biased by assigning higher eGFR estimates to patients identifying as Black, although uncertainty remains in the biological explanation underlying the race correction.3 Another model implemented in practice, the Kidney Donor Risk Index, assigns unwarrantedly higher predicted risk of kidney graft failure in patients identifying as Black, which can potentially exacerbate inequality in access to organs for transplantation.3 When a single race variable has significant potential to create bias, the likelihood of the presence of bias is much greater in black box AI models that often blindly take in a large number of variables. It is imperative, therefore, that both the developers and end users of AI-based clinical applications understand the ways in which biases arise in data and model outputs. Through this article, we aim to help the readers recognize biases in AI applications and get familiarized with methods to mitigate biases. Types of Bias in Clinical AI Applications Figure 1 illustrates the types of biases that can arise throughout different stages of AI development. At a high level, there is the algorithmic side and human side of biases, as described below.Figure 1: Stages of artificial intelligence application development and associated biases. Text in red indicates the type of bias, text in blue indicates nonalgorithmic bias mitigation, and text in tan indicates algorithmic debiasing method application.Bias in Data-Generating Process AI development begins with collecting patient data, which almost always comes from a selected sample among the underlying target population. Skewness in patient sampling can lead to disparate model performance in over- or under-represented subgroups. Differences in outcome ascertainment, such as higher sensitivity or specificity of an event of interest in electronic health data, can be another source of bias. A previous study has shown how the accurate capture of health care cost as the outcome resulted in a model preferentially recommending White patients for additional treatment resources because less money is spent on Black patients compared with White patients with similar levels of morbidity.4 Clinician bias and complex evaluation process have unfavored Black patients, leading to disparity in receiving kidney transplant; data accurately capturing this practice can generate a model that treats Black race as a risk factor for transplant failure, reinforcing the underlying inequity.5 Bias in Model Training, Testing, and Validation Often unknown to consumers of AI, numerous modeling decisions take place during the course of development. Unlike traditional medicine where publication of study protocols has become a standard practice, convoluted process of model selection, training and testing, and validation is seldom prespecified or communicated, although it has substantial effect on the outcomes. Data quality can be a function of sociodemographic factors if access to care is associated with reliable capture of data. If a group of people have a lot of missing data because of several barriers to health care, AI models will likely underperform for this group and can lead to more harm than benefits if missing data are simply excluded in model training. Similarly, various model updates that take place upon observing subpar performance, missing data treatment methods, and the decision threshold for hemodialysis can affect the performance of models predicting AKI.6 Bias in Interpretation and Application of AI Models The human side of bias plays a significant role in translation of AI to clinical benefits. Clinician trust and acceptance in AI can be a deciding factor over the actual model performance for the extent of real-world application of AI. Variability in the levels of health literacy and cultural acceptance among patients can lead to missed opportunities of improving patient outcomes through novel technologies. Importantly, the patient-provider relationship that is a product of history, culture, and mutual trust can be modified through deployment of AI in clinical nephrology, which can have unintended consequences such as loss of trust and authority or reduced adherence to medical advice.7 Ways to Mitigate and Prevent Biases in Clinical AI Application Algorithmic Debiasing Methods Biases that are algorithmic in nature, that is related to data sampling, model training, and obtaining outputs, can be addressed at least in part through the debiasing methods. In this sense, bias often refers to unwarranted statistical associations between patient attributes of interest and the outcome. Existing algorithmic debiasing methods can be categorized into preprocessing, in-processing, and postprocessing methods.8 Preprocessing methods treat the training data before model fitting to address imbalances in data. An intuitive example is the reweighing method that transforms the training data to achieve balance in groups defined by sensitive attributes of interest such as race or sex. In-processing methods modify how a model learns from data in a way that reduces the influence of a variable in the learning process. As the name implies, postprocessing takes place after a model is fitted and adjusts the outputs in a post hoc manner to address biases. Tools exist publicly to enable people to readily apply these methods in practice.8 Nonalgorithmic Bias Mitigation Completely unbiased sampling of data is usually unfeasible, so it is the responsibility of both developers and users of AI to evaluate patient representation bias. Comparing the distribution of patient demographics between training data and target population is a good starting point. Implicit bias in patient care, such as partial recommendation of novel treatment options, can be identified through examining electronic health data that reflect the practice patterns. Efforts to increase diversity in data collection and to provide equitable treatment options should accompany the aforementioned activities. In addition, detailed and transparent documentation of the modeling process, including publication of datasets and code, should become a norm in the field. It can also incentivise researchers to perform replication studies and sensitivity analyses that are critical in ascertaining clinical benefits of AI. Finally, patient and provider education is paramount to ensuring unbiased interpretation and utilization of AI. In conclusion, big data and AI utilization is an inevitable wave in medicine, and nephrology is no exception. Rigorous bias evaluation and mitigation throughout the development and application process can prevent biased AI from adversely affecting patients and health systems, especially those who are underserved. Recent efforts in providing the public with a guideline or playbook for navigating this process is important progress toward achieving fair and equitable utilization of AI.9 The epitome of AI is its ability to stay “live” and continuously learn over time, calling out the need for continuous monitoring and retraining of models to ensure unbiasedness of data and model outputs.10

科研通智能强力驱动
Strongly Powered by AbleSci AI
更新
PDF的下载单位、IP信息已删除 (2025-6-4)

科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
jerry发布了新的文献求助10
1秒前
orixero应助微笑咖啡豆采纳,获得10
3秒前
XDSH完成签到 ,获得积分10
4秒前
123完成签到,获得积分10
6秒前
shine完成签到,获得积分10
7秒前
顺利顺利完成签到 ,获得积分10
8秒前
8秒前
9秒前
9秒前
Jasper应助yx采纳,获得10
9秒前
PJT417完成签到,获得积分20
10秒前
tuanheqi应助NEKO采纳,获得50
11秒前
有趣的银完成签到,获得积分10
11秒前
飞快的孱完成签到,获得积分10
12秒前
星辰大海应助青竹采纳,获得10
12秒前
萨利发布了新的文献求助10
13秒前
Auralis完成签到 ,获得积分10
14秒前
hhh完成签到,获得积分10
14秒前
拉拉LA发布了新的文献求助10
14秒前
CodeCraft应助yatou采纳,获得10
14秒前
14秒前
Criminology34举报三七求助涉嫌违规
15秒前
老武发布了新的文献求助10
16秒前
淡淡土豆应助huayi采纳,获得10
18秒前
但行好事发布了新的文献求助10
18秒前
19秒前
科研通AI6应助贤惠的大山采纳,获得30
23秒前
sutie完成签到,获得积分10
24秒前
25秒前
26秒前
Criminology34举报哒哒哒求助涉嫌违规
27秒前
28秒前
老武完成签到,获得积分10
29秒前
huayi完成签到,获得积分10
32秒前
33秒前
小匡完成签到 ,获得积分10
35秒前
十七完成签到,获得积分10
37秒前
kdjc完成签到 ,获得积分10
38秒前
乐观夜春完成签到,获得积分10
40秒前
Sula37发布了新的文献求助30
40秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
List of 1,091 Public Pension Profiles by Region 1561
Specialist Periodical Reports - Organometallic Chemistry Organometallic Chemistry: Volume 46 1000
Current Trends in Drug Discovery, Development and Delivery (CTD4-2022) 800
Foregrounding Marking Shift in Sundanese Written Narrative Segments 600
Holistic Discourse Analysis 600
Beyond the sentence: discourse and sentential form / edited by Jessica R. Wirth 600
热门求助领域 (近24小时)
化学 材料科学 医学 生物 工程类 有机化学 生物化学 物理 纳米技术 计算机科学 内科学 化学工程 复合材料 物理化学 基因 遗传学 催化作用 冶金 量子力学 光电子学
热门帖子
关注 科研通微信公众号,转发送积分 5522356
求助须知:如何正确求助?哪些是违规求助? 4613399
关于积分的说明 14538602
捐赠科研通 4551100
什么是DOI,文献DOI怎么找? 2493968
邀请新用户注册赠送积分活动 1475030
关于科研通互助平台的介绍 1446408