离群值
数学
主成分分析
算法
数学优化
缩小
最大化
规范(哲学)
数据点
矩阵范数
降维
计算机科学
模式识别(心理学)
人工智能
特征向量
统计
物理
量子力学
法学
政治学
作者
Minnan Luo,Feiping Nie,Xiaojun Chang,Yi Yang,Alexander G. Hauptmann,Qinghua Zheng
摘要
Robust principal component analysis (PCA) is one of the most important dimension-reduction techniques for handling high-dimensional data with outliers. However, most of the existing robust PCA presupposes that the mean of the data is zero and incorrectly utilizes the average of data as the optimal mean of robust PCA. In fact, this assumption holds only for the squared [Formula: see text]-norm-based traditional PCA. In this letter, we equivalently reformulate the objective of conventional PCA and learn the optimal projection directions by maximizing the sum of projected difference between each pair of instances based on [Formula: see text]-norm. The proposed method is robust to outliers and also invariant to rotation. More important, the reformulated objective not only automatically avoids the calculation of optimal mean and makes the assumption of centered data unnecessary, but also theoretically connects to the minimization of reconstruction error. To solve the proposed nonsmooth problem, we exploit an efficient optimization algorithm to soften the contributions from outliers by reweighting each data point iteratively. We theoretically analyze the convergence and computational complexity of the proposed algorithm. Extensive experimental results on several benchmark data sets illustrate the effectiveness and superiority of the proposed method.
科研通智能强力驱动
Strongly Powered by AbleSci AI