可解释性
MNIST数据库
计算机科学
人工神经网络
特征(语言学)
编码(集合论)
深层神经网络
人工智能
黑匣子
反向传播
深度学习
模式识别(心理学)
机器学习
集合(抽象数据类型)
程序设计语言
哲学
语言学
作者
Avanti Shrikumar,Peyton G. Greenside,Anshul Kundaje
出处
期刊:International Conference on Machine Learning
日期:2017-08-06
卷期号:: 3145-3153
被引量:617
摘要
The purported black box nature of neural networks is a barrier to adoption in applications where interpretability is essential. Here we present DeepLIFT (Deep Learning Important FeaTures), a method for decomposing the output prediction of a neural network on a specific input by backpropagating the contributions of all neurons in the network to every feature of the input. DeepLIFT compares the activation of each neuron to its 'reference activation' and assigns contribution scores according to the difference. By optionally giving separate consideration to positive and negative contributions, DeepLIFT can also reveal dependencies which are missed by other approaches. Scores can be computed efficiently in a single backward pass. We apply DeepLIFT to models trained on MNIST and simulated genomic data, and show significant advantages over gradient-based methods. Video tutorial: http://goo.gl/qKb7pL, code: http://goo.gl/RM8jvH.
科研通智能强力驱动
Strongly Powered by AbleSci AI