计算机科学
人工智能
卷积神经网络
透视图(图形)
分割
聚类分析
点(几何)
机器学习
人工神经网络
可靠性(半导体)
钥匙(锁)
深度学习
图像分割
功率(物理)
几何学
物理
量子力学
计算机安全
数学
作者
Seonggyeom Kim,Dong‐Kyu Chae
标识
DOI:10.1109/tpami.2024.3357717
摘要
Model explainability is one of the crucial ingredients for building trustable AI systems, especially in the applications requiring reliability such as automated driving and diagnosis. Many explainability methods have been studied in the literature. Among many others, this article focuses on a research line that tries to visually explain a pre-trained image classification model such as Convolutional Neural Network by discovering concepts learned by the model, which is so-called the concept-based explanation. Previous concept-based explanation methods rely on the human definition of concepts (e.g., the Broden dataset) or semantic segmentation techniques like Slic (Simple Linear Iterative Clustering). However, we argue that the concepts identified by those methods may show image parts which are more in line with a human perspective or cropped by a segmentation method, rather than purely reflect a model's own perspective. We propose Model-Oriented Concept Extraction (MOCE), a novel approach to extracting key concepts based solely on a model itself, thereby being able to capture its unique perspectives which are not affected by any external factors. Experimental results on various pre-trained models confirmed the advantages of extracting concepts by truly representing the model's point of view.
科研通智能强力驱动
Strongly Powered by AbleSci AI