自编码
人工智能
非线性降维
特征学习
外部数据表示
计算机科学
代表(政治)
正规化(语言学)
嵌入
核(代数)
模式识别(心理学)
歧管对齐
歧管(流体力学)
核方法
深度学习
可逆矩阵
机器学习
数学
降维
支持向量机
纯数学
组合数学
法学
工程类
政治
机械工程
政治学
作者
Andrés Duque,Sacha Morin,Guy Wolf,Kevin R. Moon
标识
DOI:10.1109/tpami.2022.3222104
摘要
A fundamental task in data exploration is to extract low dimensional representations that capture intrinsic geometry in data, especially for faithfully visualizing data in two or three dimensions. Common approaches use kernel methods for manifold learning. However, these methods typically only provide an embedding of the input data and cannot extend naturally to new data points. Autoencoders have also become popular for representation learning. While they naturally compute feature extractors that are extendable to new data and invertible (i.e., reconstructing original features from latent representation), they often fail at representing the intrinsic data geometry compared to kernel-based manifold learning. We present a new method for integrating both approaches by incorporating a geometric regularization term in the bottleneck of the autoencoder. This regularization encourages the learned latent representation to follow the intrinsic data geometry, similar to manifold learning algorithms, while still enabling faithful extension to new data and preserving invertibility. We compare our approach to autoencoder models for manifold learning to provide qualitative and quantitative evidence of our advantages in preserving intrinsic structure, out of sample extension, and reconstruction. Our method is easily implemented for big-data applications, whereas other methods are limited in this regard.
科研通智能强力驱动
Strongly Powered by AbleSci AI