矩阵范数
缩小
计算机科学
趋同(经济学)
低秩近似
规范(哲学)
人工智能
数学优化
秩(图论)
基质(化学分析)
算法
数学
特征向量
组合数学
量子力学
物理
数学分析
经济增长
复合材料
经济
汉克尔矩阵
材料科学
法学
政治学
作者
Zhi Wang,Dong Hu,Zhuo Liu,Chao Gao,Zheng Wang
标识
DOI:10.1109/tpami.2024.3512458
摘要
In recent years, a large number of studies have shown that low rank matrix learning (LRML) has become a popular approach in machine learning and computer vision with many important applications, such as image inpainting, subspace clustering, and recommendation system. The latest LRML methods resort to using some surrogate functions as convex or nonconvex relaxation of the rank function. However, most of these methods ignore the difference between different rank components and can only yield suboptimal solutions. To alleviate this problem, in this paper we propose a novel nonconvex regularizer called capped reweighting norm minimization (CRNM), which not only considers the different contributions of different rank components, but also adaptively truncates sequential singular values. With it, a general LRML model is obtained. Meanwhile, under some mild conditions, the global optimum of CRNM regularized least squares subproblem can be easily obtained in closed-form. Through the analysis of the theoretical properties of CRNM, we develop a high computational efficiency optimization method with convergence guarantee to solve the general LRML model. More importantly, by using the Kurdyka-Łojasiewicz (KŁ) inequality, its local and global convergence properties are established. Finally, we show that the proposed nonconvex regularizer as well as the optimization approach are suitable for different low rank tasks, such as matrix completion and subspace clustering. Extensive experimental results demonstrate that the constructed models and methods provide significant advantages over several state-of-the-art low rank matrix leaning models and methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI