首页> 外文会议>IEEE Global Communications Conference >Multi-Agent Reinforcement Learning Enabling Dynamic Pricing Policy for Charging Station Operators
【24h】

Multi-Agent Reinforcement Learning Enabling Dynamic Pricing Policy for Charging Station Operators

机译:支持充电站运营商动态定价策略的多智能体强化学习

获取原文

摘要

The development of plug-in electric vehicles (PEVs) brings lucrative opportunities for charging station operators (CSOs). To attract more CSOs to the PEV market, provision of reasonable pricing policy is of great importance. However, dynamic environments and uncertain behavior of competitors make the pricing problem of CSOs challenging. In this paper, we focus on the dynamic pricing policy for maximizing the long-term profits of CSOs. Firstly, we propose a hierarchical framework to describe the economic association of PEV market, which is composed of smart grid, CSOs and charging stations (CSs) serving PEVs from top to bottom. Next, we leverage the Markov game to model the layer of CSOs as a competitive market. Finally, we design a dynamic pricing policy algorithm (DPPA) based on multi-agent reinforcement learning to achieve higher long-term profits of CSOs. Based on the real data of PEVs in Beijing, the experiment results show that DPPA has a significant improvement in long-term profit of CSOs, and the improvement gains increase over time. Moreover, DPPA can reduce the profit loss of CSOs effectively while involving more competitors.
机译:插入电动车辆(PEVS)的开发为充电站运营商(CSOS)带来了利润丰厚的机会。为了吸引更多的CSO到PEV市场,提供合理的定价政策具有重要意义。然而,动态环境和竞争对手的不确定行为使CSOS挑战的定价问题。在本文中,我们专注于最大化CSO的长期利润的动态定价政策。首先,我们提出了一个分层框架来描述PEV市场的经济协会,它由智能电网,CSOS和充电站(CSS)从顶部到底部提供PEV。接下来,我们利用马尔可夫游戏将CSO层层模拟为竞争市场。最后,我们设计了一种基于多功能钢筋学习的动态定价策略算法(DPPA),以实现CSO的较高长期利润。基于北京的PEV的真实数据,实验结果表明,DPPA对CSO的长期利润有显着改善,随着时间的推移,提升的提升增加。此外,DPPA可以有效地减少CSO的利润损失,同时涉及更多竞争对手。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号