首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Deep Decentralized Multi-task Multi-Agent Reinforcement Learning under Partial Observability
【24h】

Deep Decentralized Multi-task Multi-Agent Reinforcement Learning under Partial Observability

机译:部分可观察性下的深度分散多任务多智能体强化学习

获取原文
       

摘要

Many real-world tasks involve multiple agents with partial observability and limited communication. Learning is challenging in these settings due to local viewpoints of agents, which perceive the world as non-stationary due to concurrently-exploring teammates. Approaches that learn specialized policies for individual tasks face problems when applied to the real world: not only do agents have to learn and store distinct policies for each task, but in practice identities of tasks are often non-observable, making these approaches inapplicable. This paper formalizes and addresses the problem of multi-task multi-agent reinforcement learning under partial observability. We introduce a decentralized single-task learning approach that is robust to concurrent interactions of teammates, and present an approach for distilling single-task policies into a unified policy that performs well across multiple related tasks, without explicit provision of task identity.
机译:许多实际任务涉及多个具有部分可观察性且通信受限的代理。在这些环境中,由于特工的局部观点,在这些环境中学习颇具挑战性,由于同时探索队友,他们认为世界是不稳定的。学习针对单个任务的专用策略的方法在应用于现实世界时会遇到问题:不仅代理必须为每个任务学习和存储不同的策略,而且在实践中,任务的身份通常不可观察,从而使这些方法不适用。本文对部分可观察性下的多任务多智能体强化学习问题进行了形式化处理。我们介绍了一种分散的单任务学习方法,该方法对队友的并发交互具有鲁棒性,并提出了一种方法,用于将单任务策略提炼为可在多个相关任务之间良好执行的统一策略,而无需明确提供任务身份。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号