首页> 外文会议>Annual Conference on Information Sciences and Systems >Escaping Saddle Points for Zeroth-order Non-convex Optimization using Estimated Gradient Descent
【24h】

Escaping Saddle Points for Zeroth-order Non-convex Optimization using Estimated Gradient Descent

机译:使用估计的梯度下降为零阶非凸优化转义鞍点

获取原文

摘要

Gradient descent and its variants are widely used in machine learning. However, oracle access of gradient may not be available in many applications, limiting the direct use of gradient descent. This paper proposes a method of estimating gradient to perform gradient descent, that converges to a second-order stationary point for general non-convex optimization problems. Beyond the first-order stationary properties, the second-order stationary properties are important in machine learning applications to achieve better performance. We show that the proposed model-free non-convex optimization algorithm returns an ε-second-order stationary point with $ilde Oleft( {rac{{{d^{2 + rac{heta }{2}}}}}{{{arepsilon ^{8 + heta }}}}} ight)$ queries of the function for any arbitrary θ > 0.
机译:梯度下降及其变体广泛用于机器学习中。但是,在许多应用程序中可能无法使用oracle进行梯度访问,从而限制了直接使用梯度下降。本文提出了一种估计梯度以进行梯度下降的方法,该方法收敛到一般非凸优化问题的二阶固定点。除了一阶平稳特性外,二阶平稳特性在机器学习应用程序中对于实现更好的性能也很重要。我们表明,所提出的无模型非凸优化算法返回了带有$ \ tilde O \ left({\ frac {{{d ^ {2 + \ frac {\ theta} {2} }}}} {{{{varepsilon ^ {8 + \ theta}}}}} \ right)$查询任意θ> 0的函数。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号