李雅普诺夫函数
数学
趋同(经济学)
常微分方程
加速度
应用数学
收敛速度
凸函数
数学优化
算法
随机微分方程
微分方程
正多边形
计算机科学
数学分析
钥匙(锁)
非线性系统
经济增长
量子力学
经典力学
物理
计算机安全
经济
几何学
作者
Paul Dobson,J. M. Sanz‐Serna,Konstantinos C. Zygalakis
出处
期刊:Cornell University - arXiv
日期:2023-01-01
标识
DOI:10.48550/arxiv.2305.08658
摘要
We revisit the general framework introduced by Fazylab et al. (SIAM J. Optim. 28, 2018) to construct Lyapunov functions for optimization algorithms in discrete and continuous time. For smooth, strongly convex objective functions, we relax the requirements necessary for such a construction. As a result we are able to prove for Polyak's ordinary differential equations and for a two-parameter family of Nesterov algorithms rates of convergence that improve on those available in the literature. We analyse the interpretation of Nesterov algorithms as discretizations of the Polyak equation. We show that the algorithms are instances of Additive Runge-Kutta integrators and discuss the reasons why most discretizations of the differential equation do not result in optimization algorithms with acceleration. We also introduce a modification of Polyak's equation and study its convergence properties. Finally we extend the general framework to the stochastic scenario and consider an application to random algorithms with acceleration for overparameterized models; again we are able to prove convergence rates that improve on those in the literature.
科研通智能强力驱动
Strongly Powered by AbleSci AI