首页> 外文会议>International Conference on Robotics and Automation >Continuous Value Iteration (CVI) Reinforcement Learning and Imaginary Experience Replay (IER) For Learning Multi-Goal, Continuous Action and State Space Controllers
【24h】

Continuous Value Iteration (CVI) Reinforcement Learning and Imaginary Experience Replay (IER) For Learning Multi-Goal, Continuous Action and State Space Controllers

机译:用于学习多目标,连续动作和状态空间控制器的连续值迭代(CVI)强化学习和虚幻体验重放(IER)

获取原文

摘要

This paper presents a novel model-free Reinforcement Learning algorithm for learning behavior in continuous action, state, and goal spaces. The algorithm approximates optimal value functions using non-parametric estimators. It is able to efficiently learn to reach multiple arbitrary goals in deterministic and nondeterministic environments. To improve generalization in the goal space, we propose a novel sample augmentation technique. Using these methods, robots learn faster and overall better controllers. We benchmark the proposed algorithms using simulation and a real-world voltage controlled robot that learns to maneuver in a non-observable Cartesian task space.
机译:本文提出了一种新颖的无模型强化学习算法,用于学习连续动作,状态和目标空间中的行为。该算法使用非参数估计量来近似最佳值函数。它能够有效地学习在确定性和非确定性环境中达到多个任意目标。为了提高目标空间的概括性,我们提出了一种新颖的样本增强技术。使用这些方法,机器人可以学习更快,整体上更好的控制器。我们使用仿真和在不可见的笛卡尔任务空间中进行机动的真实世界电压控制机器人对提出的算法进行基准测试。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号