首页> 外文会议>Annual conference on Neural Information Processing Systems >Accelerating Stochastic Gradient Descent using Predictive Variance Reduction
【24h】

Accelerating Stochastic Gradient Descent using Predictive Variance Reduction

机译:使用预测方差减少加速随机梯度下降

获取原文

摘要

Stochastic gradient descent is popular for large scale optimization but has slow convergence asymptotically due to the inherent variance. To remedy this problem, we introduce an explicit variance reduction method for stochastic gradient descent which we call stochastic variance reduced gradient (SVRG). For smooth and strongly convex functions, we prove that this method enjoys the same fast convergence rate as those of stochastic dual coordinate ascent (SDCA) and Stochastic Average Gradient (SAG). However, our analysis is significantly simpler and more intuitive. Moreover, unlike SDCA or SAG, our method does not require the storage of gradients, and thus is more easily applicable to complex problems such as some structured prediction problems and neural network learning.
机译:随机梯度下降是大规模优化的流行,但由于固有的差异而渐近的收敛性慢。为了解决这个问题,我们介绍了一种用于随机梯度下降的显式方差减少方法,其呼叫随机方差减少梯度(SVRG)。对于光滑和强大的凸起功能,我们证明该方法享有与随机双坐标上升(SDCA)和随机平均梯度(SAG)相同的快速收敛速度。但是,我们的分析明显更简单,更直观。此外,与SDCA或SAG不同,我们的方法不需要存储梯度,因此更容易适用于一些结构化预测问题和神经网络学习等复杂问题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号