方差减少
随机梯度下降算法
趋同(经济学)
差异(会计)
计算机科学
数学优化
梯度下降
对偶(语法数字)
收敛速度
随机优化
凸函数
应用数学
数学
正多边形
人工神经网络
人工智能
艺术
计算机网络
频道(广播)
几何学
会计
文学类
经济
业务
经济增长
作者
Rie Johnson,Tong Zhang
摘要
Stochastic gradient descent is popular for large scale optimization but has slow convergence asymptotically due to the inherent variance. To remedy this problem, we introduce an explicit variance reduction method for stochastic gradient descent which we call stochastic variance reduced gradient (SVRG). For smooth and strongly convex functions, we prove that this method enjoys the same fast convergence rate as those of stochastic dual coordinate ascent (SDCA) and Stochastic Average Gradient (SAG). However, our analysis is significantly simpler and more intuitive. Moreover, unlike SDCA or SAG, our method does not require the storage of gradients, and thus is more easily applicable to complex problems such as some structured prediction problems and neural network learning.
科研通智能强力驱动
Strongly Powered by AbleSci AI