计算机科学
等价(形式语言)
归属
人工神经网络
人工智能
深层神经网络
公制(单位)
透视图(图形)
构造(python库)
算法
机器学习
理论计算机科学
数学
心理学
社会心理学
运营管理
经济
程序设计语言
离散数学
作者
Marco Ancona,Enea Ceolini,Cengiz Öztireli,Markus Groß
出处
期刊:Cornell University - arXiv
日期:2017-01-01
被引量:264
标识
DOI:10.48550/arxiv.1711.06104
摘要
Understanding the flow of information in Deep Neural Networks (DNNs) is a challenging problem that has gain increasing attention over the last few years. While several methods have been proposed to explain network predictions, there have been only a few attempts to compare them from a theoretical perspective. What is more, no exhaustive empirical comparison has been performed in the past. In this work, we analyze four gradient-based attribution methods and formally prove conditions of equivalence and approximation between them. By reformulating two of these methods, we construct a unified framework which enables a direct comparison, as well as an easier implementation. Finally, we propose a novel evaluation metric, called Sensitivity-n and test the gradient-based attribution methods alongside with a simple perturbation-based attribution method on several datasets in the domains of image and text classification, using various network architectures.
科研通智能强力驱动
Strongly Powered by AbleSci AI