可解释性
机器学习
人气
计算机科学
树(集合论)
人工智能
集合(抽象数据类型)
决策树
随机森林
人口
特征(语言学)
数学
医学
哲学
数学分析
环境卫生
程序设计语言
社会心理学
语言学
心理学
作者
Scott Lundberg,Gabriel Erion,Hugh Chen,Alex J. DeGrave,Jordan M. Prutkin,Bala G. Nair,Ronit Katz,Jonathan Himmelfarb,Nisha Bansal,Su‐In Lee
标识
DOI:10.1038/s42256-019-0138-9
摘要
Tree-based machine learning models such as random forests, decision trees and gradient boosted trees are popular nonlinear predictive models, yet comparatively little attention has been paid to explaining their predictions. Here we improve the interpretability of tree-based models through three main contributions. (1) A polynomial time algorithm to compute optimal explanations based on game theory. (2) A new type of explanation that directly measures local feature interaction effects. (3) A new set of tools for understanding global model structure based on combining many local explanations of each prediction. We apply these tools to three medical machine learning problems and show how combining many high-quality local explanations allows us to represent global structure while retaining local faithfulness to the original model. These tools enable us to (1) identify high-magnitude but low-frequency nonlinear mortality risk factors in the US population, (2) highlight distinct population subgroups with shared risk characteristics, (3) identify nonlinear interaction effects among risk factors for chronic kidney disease and (4) monitor a machine learning model deployed in a hospital by identifying which features are degrading the model’s performance over time. Given the popularity of tree-based machine learning models, these improvements to their interpretability have implications across a broad set of domains. Tree-based machine learning models are widely used in domains such as healthcare, finance and public services. The authors present an explanation method for trees that enables the computation of optimal local explanations for individual predictions, and demonstrate their method on three medical datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI