首页> 外文会议>AAAI Conference on Artificial Intelligence >A Stochastic Derivative-Free Optimization Method with Importance Sampling: Theory and Learning to Control
【24h】

A Stochastic Derivative-Free Optimization Method with Importance Sampling: Theory and Learning to Control

机译:具有重要性抽样的随机衍生优化方法:理论和学习控制

获取原文

摘要

We consider the problem of unconstrained minimization of a smooth objective function in R~n in a setting where only function evaluations are possible. While importance sampling is one of the most popular techniques used by machine learning practitioners to accelerate the convergence of their models when applicable, there is not much existing theory for this acceleration in the derivative-free setting. In this paper, we propose the first derivative free optimization method with importance sampling and derive new improved complexity results on non-convex, convex and strongly convex functions. We conduct extensive experiments on various synthetic and real LIBSVM datasets confirming our theoretical results. We test our method on a collection of continuous control tasks on MuJoCo environments with varying difficulty. Experiments show that our algorithm is practical for high dimensional continuous control problems where importance sampling results in a significant sample complexity improvement.
机译:我们考虑在只有功能评估的情况下,我们在r〜n中对r〜n中的平滑目标函数的不受约束最小化的问题。虽然重要性采样是机器学习从业者使用的最受欢迎的技术之一,用于在适用时加速其模型的收敛性,但在无衍生物环境中的加速度没有太大的现有理论。在本文中,我们提出了具有重要性采样的第一种衍生自由优化方法,并导出了非凸,凸强和强凸函数的新改进的复杂性结果。我们对各种合成和真正的LIBSVM数据集进行了广泛的实验,确认了我们的理论结果。我们在Mujoco环境中的连续控制任务集合中测试了我们的方法,具有不同的难度。实验表明,我们的算法对于高维连续控制问题实用,重要的采样导致显着的样本复杂性改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号