计算机科学
随机梯度下降算法
共轭梯度法
水准点(测量)
数学优化
梯度下降
机器学习
算法
人工智能
数学
人工神经网络
大地测量学
地理
出处
期刊:IEEE Transactions on Big Data
[Institute of Electrical and Electronics Engineers]
日期:2023-08-01
卷期号:9 (6): 1598-1606
被引量:2
标识
DOI:10.1109/tbdata.2023.3300546
摘要
The extreme success of stochastic optimization (SO) in large-scale machine learning problems, information retrieval, bioinformatics, etc., has been widely reported, especially in recent years. As an effective tactic, conjugate gradient (CG) has been gaining its popularity in accelerating SO algorithms. This paper develops a novel type of stochastic conjugate gradient descent (SCG) algorithms from the perspective of the Powerball strategy and the hypergradient descent (HD) technique. The crucial idea behind the resulting methods is inspired by pursuing the equilibrium of ordinary differential equations (ODEs). We elucidate the effect of the Powerball strategy in SCG algorithms. The introduction of HD, on the other side, makes the resulting methods work with an online learning rate. Meanwhile, we provide a comprehension of the theoretical results for the resulting algorithms under non-convex assumptions. As a byproduct, we bridge the gap between the learning rate and powered stochastic optimization (PSO) algorithms, which is still an open problem. Resorting to numerical experiments on numerous benchmark datasets, we test the parameter sensitivity of the proposed methods and demonstrate the superior performance of our new algorithms over state-of-the-art algorithms.
科研通智能强力驱动
Strongly Powered by AbleSci AI