...
首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Mini-batch Block-coordinate based Stochastic Average Adjusted Gradient Methods to Solve Big Data Problems
【24h】

Mini-batch Block-coordinate based Stochastic Average Adjusted Gradient Methods to Solve Big Data Problems

机译:基于迷你批处理块坐标的随机平均调整后的梯度方法来解决大数据问题

获取原文
           

摘要

Big Data problems in Machine Learning have large number of data points or large number of features, or both, which make training of models difficult because of high computational complexities of single iteration of learning algorithms. To solve such learning problems, Stochastic Approximation offers an optimization approach to make complexity of each iteration independent of number of data points by taking only one data point or mini-batch of data points during each iteration and thereby helping to solve problems with large number of data points. Similarly, Coordinate Descent offers another optimization approach to make iteration complexity independent of the number of features/coordinates/variables by taking only one feature or block of features, instead of all, during an iteration and thereby helping to solve problems with large number of features. In this paper, an optimization framework, namely, Batch Block Optimization Framework has been developed to solve big data problems using the best of Stochastic Approximation as well as the best of Coordinate Descent approaches, independent of any solver. This framework is used to solve strongly convex and smooth empirical risk minimization problem with gradient descent (as a solver) and two novel Stochastic Average Adjusted Gradient methods have been proposed to reduce variance in mini-batch and block-coordinate setting of the developed framework. Theoretical analysis prove linear convergence of the proposed methods and empirical results with bench marked datasets prove the superiority of proposed methods against existing methods.
机译:机器学习中的大数据问题具有大量数据点或大量功能,或两者都是由于学习算法的单一迭代的高计算复杂性而难以实现模型的训练。为了解决这种学习问题,随机近似提供了一种优化方法,可以通过在每次迭代期间仅在每个迭代期间仅在每个数据点或微型数据点中独立于数据点的数量而无关的每个迭代的复杂性。从而有助于解决大量问题的问题数据点。类似地,坐标血压提供了另一种优化方法,可以通过仅在迭代期间仅通过仅取迭代或特征块而不是所有特征或特征来进行迭代复杂性,而不是所有特征,而不是迭代,从而有助于解决大量特征的问题。在本文中,已经开发了一种优化框架,即,使用最佳随机近似以及独立于任何求解器的坐标血压方法来解决大数据问题来解决大数据问题。该框架用于解决强烈凸起和平稳的经验风险最小化问题,梯度下降(作为求解器),已经提出了两种新的随机平均调节梯度方法,以降低开发框架的迷你批量和块坐标设置中的方差。理论分析证明了所提出的方法的线性融合和具有基准的实证结果,标记数据集证明了对现有方法的提出方法的优越性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号