首页> 外文期刊>IEICE Transactions on Information and Systems >Model-Based Reinforcement Learning in Multiagent Systems with Sequential Action Selection
【24h】

Model-Based Reinforcement Learning in Multiagent Systems with Sequential Action Selection

机译:具有顺序动作选择的多主体系统中基于模型的强化学习

获取原文
获取原文并翻译 | 示例
           

摘要

Model-based reinforcement learning uses the gathered information, during each experience, more efficiently than model-free reinforcement learning. This is especially interesting in multiagent systems, since a large number of experiences are necessary to achieve a good performance. In this paper, model-based reinforcement learning is developed for a group of self-interested agents with sequential action selection based on traditional prioritized sweeping. Every single situation of decision making in this learning process, called extensive Markov game, is modeled as n-person general-sum extensive form game with perfect information. A modified version of backward induction is proposed for action selection, which adjusts the tradeoff between selecting subgame perfect equilibrium points, as the optimal joint actions, and learning new joint actions. The algorithm is proved to be convergent and discussed based on the new results on the convergence of the traditional prioritized sweeping.
机译:与无模型的强化学习相比,基于模型的强化学习在每次体验期间更有效地使用所收集的信息。在多代理系统中,这尤其有趣,因为要获得良好的性能,需要大量的经验。在本文中,基于模型的强化学习是针对一组自私的代理开发的,这些代理具有基于传统优先清扫的顺序动作选择。在这种学习过程中,决策的每一个情况,称为广泛马尔可夫博弈,都被建模为具有完善信息的n人一般和广义博弈。提出了一种改进的后向归纳法用于动作选择,它可以调整选择子游戏的完美平衡点(作为最佳联合动作)与学习新的联合动作之间的权衡。在传统优先扫描的收敛性的新结果的基础上,证明了该算法是收敛的并进行了讨论。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号