人工智能
计算机科学
机器学习
贝叶斯概率
最大后验估计
概率逻辑
对象(语法)
刮擦
模式识别(心理学)
目标检测
视觉对象识别的认知神经科学
计算机视觉
卷积神经网络
判别式
特征(语言学)
任务(项目管理)
深度学习
特征提取
数学
最大似然
统计
操作系统
作者
Feifei Li,Rob Fergus,Pietro Perona
标识
DOI:10.1109/tpami.2006.79
摘要
Learning visual models of object categories notoriously requires hundreds or thousands of training examples. We show that it is possible to learn much information about a category from just one, or a handful, of images. The key insight is that, rather than learning from scratch, one can take advantage of knowledge coming from previously learned categories, no matter how different these categories might be. We explore a Bayesian implementation of this idea. Object categories are represented by probabilistic models. Prior knowledge is represented as a probability density function on the parameters of these models. The posterior model for an object category is obtained by updating the prior in the light of one or more observations. We test a simple implementation of our algorithm on a database of 101 diverse object categories. We compare category models learned by an implementation of our Bayesian approach to models learned from by maximum likelihood (ML) and maximum a posteriori (MAP) methods. We find that on a database of more than 100 categories, the Bayesian approach produces informative models when the number of training examples is too small for other methods to operate successfully.
科研通智能强力驱动
Strongly Powered by AbleSci AI