Softmax函数
计算机科学
MNIST数据库
人工智能
机器学习
公制(单位)
集合(抽象数据类型)
阈值
面子(社会学概念)
熵(时间箭头)
班级(哲学)
深度学习
社会学
物理
经济
图像(数学)
量子力学
社会科学
程序设计语言
运营管理
作者
Akshay Raj Dhamija,Manuel Günther,Terrance E. Boult
出处
期刊:Neural Information Processing Systems
日期:2018-01-01
卷期号:31: 9157-9168
被引量:92
摘要
Agnostophobia, the fear of the unknown, can be experienced by deep learning engineers while applying their networks to real-world applications. Unfortunately, network behavior is not well defined for inputs far from a networks training set. In an uncontrolled environment, networks face many instances that are not of interest to them and have to be rejected in order to avoid a false positive. This problem has previously been tackled by researchers by either a) thresholding softmax, which by construction cannot return none of the known classes, or b) using an additional background or garbage class. In this paper, we show that both of these approaches help, but are generally insufficient when previously unseen classes are encountered. We also introduce a new evaluation metric that focuses on comparing the performance of multiple approaches in scenarios where such unseen classes or unknowns are encountered. Our major contributions are simple yet effective Entropic Open-Set and Objectosphere losses that train networks using negative samples from some classes. These novel losses are designed to maximize entropy for unknown inputs while increasing separation in deep feature space by modifying magnitudes of known and unknown samples. Experiments on networks trained to classify classes from MNIST and CIFAR-10 show that our novel loss functions are significantly better at dealing with unknown inputs from datasets such as Devanagari, NotMNIST, CIFAR-100 and SVHN.
科研通智能强力驱动
Strongly Powered by AbleSci AI