可解释性
次线性函数
近端梯度法
计算机科学
正规化(语言学)
数学优化
一般化
正多边形
收敛速度
趋同(经济学)
最优化问题
序列(生物学)
凸优化
算法
数学
人工智能
离散数学
几何学
计算机网络
经济
数学分析
频道(广播)
生物
遗传学
经济增长
作者
Shubao Zhang,Hui Qian,Xiaojin Gong
出处
期刊:Proceedings of the ... AAAI Conference on Artificial Intelligence
[Association for the Advancement of Artificial Intelligence (AAAI)]
日期:2016-03-02
卷期号:30 (1)
被引量:7
标识
DOI:10.1609/aaai.v30i1.10253
摘要
In many learning tasks with structural properties, structured sparse modeling usually leads to better interpretability and higher generalization performance. While great efforts have focused on the convex regularization, recent studies show that nonconvex regularizers can outperform their convex counterparts in many situations. However, the resulting nonconvex optimization problems are still challenging, especially for the structured sparsity-inducing regularizers. In this paper, we propose a splitting method for solving nonconvex structured sparsity optimization problems. The proposed method alternates between a gradient step and an easily solvable proximal step, and thus enjoys low per-iteration computational complexity. We prove that the whole sequence generated by the proposed method converges to a critical point with at least sublinear convergence rate, relying on the Kurdyka-Łojasiewicz inequality. Experiments on both simulated and real-world data sets demonstrate the efficiency and efficacy of the proposed method.
科研通智能强力驱动
Strongly Powered by AbleSci AI