嵌入
概率逻辑
降维
高斯分布
人工智能
计算机科学
成对比较
对象(语法)
模式识别(心理学)
维数之咒
概率分布
数学
统计
量子力学
物理
作者
Geoffrey E. Hinton,Sam T. Roweis
出处
期刊:Neural Information Processing Systems
日期:2002-01-01
卷期号:15: 857-864
被引量:1432
摘要
We describe a probabilistic approach to the task of placing objects, described by high-dimensional vectors or by pairwise dissimilarities, in a low-dimensional space in a way that preserves neighbor identities. A Gaussian is centered on each object in the high-dimensional space and the densities under this Gaussian (or the given dissimilarities) are used to define a probability distribution over all the potential neighbors of the object. The aim of the embedding is to approximate this distribution as well as possible when the same operation is performed on the low-dimensional of the objects. A natural cost function is a sum of Kullback-Leibler divergences, one per object, which leads to a simple gradient for adjusting the positions of the low-dimensional images. Unlike other dimensionality reduction methods, this probabilistic framework makes it easy to represent each object by a mixture of widely separated low-dimensional images. This allows ambiguous objects, like the document count vector for the word bank, to have versions close to the images of both river and finance without forcing the images of outdoor concepts to be located close to those of corporate concepts.
科研通智能强力驱动
Strongly Powered by AbleSci AI