首页> 外文会议>IEEE Conference on Games >Optimal Use of Experience in First Person Shooter Environments
【24h】

Optimal Use of Experience in First Person Shooter Environments

机译:在第一人称射击环境中最佳使用经验

获取原文

摘要

Although reinforcement learning has made great strides recently, a continuing limitation is that it requires an extremely high number of interactions with the environment. In this paper, we explore the effectiveness of reusing experience from the experience replay buffer in the Deep Q-Learning algorithm. We test the effectiveness of applying learning update steps multiple times per environmental step in the VizDoom environment and show first, this requires a change in the learning rate, and second that it does not improve the performance of the agent. Furthermore, we show that updating less frequently is effective up to a ratio of 4:1, after which performance degrades significantly. These results quantitatively confirm the widespread practice of performing learning updates every 4th environmental step.
机译:尽管强化学习最近取得了长足的进步,但持续的局限性在于它要求与环境进行大量交互。在本文中,我们从深度Q学习算法中的经验重播缓冲区中探索重用经验的有效性。我们测试了在VizDoom环境中每个环境步骤多次应用学习更新步骤的有效性,并且首先表明,这需要改变学习率,其次,它不能提高代理的性能。此外,我们表明,不频繁更新最多可达到4:1的比率,此后性能会显着下降。这些结果定量地证实了每四个环境步骤进行一次学习更新的广泛实践。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号