...
首页> 外文期刊>IEEE Transactions on Cognitive Communications and Networking >A Reinforcement Learning Method for Joint Mode Selection and Power Adaptation in the V2V Communication Network in 5G
【24h】

A Reinforcement Learning Method for Joint Mode Selection and Power Adaptation in the V2V Communication Network in 5G

机译:5G中V2V通信网络中的联合模式选择和功率适应的加强学习方法

获取原文
获取原文并翻译 | 示例
           

摘要

A 5G network is the key driving factor in the development of vehicle-to-vehicle (V2V) communication technology, and V2V communication in 5G has recently attracted great interest. In the V2V communication network, users can choose different transmission modes and power levels for communication, to guarantee their quality-of-service (QoS), high capacity of vehicle-to-infrastructure (V2I) links and ultra-reliability of V2Vlinks. Aiming atV2V communication mode selection and power adaptation in 5G communication networks, a reinforcement learning (RL) framework based on slow fading parameters and statistical information is proposed. In this paper, our objective is to maximize the total capacity of V2I links while guaranteeing the strict transmission delay and reliability constraints of V2V links. Considering the fast channel variations and the continuous-valued state in a high mobility vehicular environment, we use a multi-agent double deep Q-learning (DDQN) algorithm. Each V2V link is considered as an agent, learning the optimal policy with the updated Q-network by interacting with the environment. Experiments verify the convergence of our algorithm. The simulation results show that the proposed scheme can significantly optimize the total capacity of the V2I links and ensure the latency and reliability requirements of the V2V links.
机译:5G网络是车辆到车辆(V2V)通信技术开发的关键驱动因素,而5G的V2V通信最近引起了极大的兴趣。在V2V通信网络中,用户可以选择不同的传输模式和电源电平进行通信,以保证它们的服务质量(QoS),高容量的车辆到基础设施(V2I)链路和V2Vlink的超可靠性。提出了一种ATV2V通信模式选择和功率适应5G通信网络,提出了一种基于慢衰落参数和统计信息的加强学习(RL)框架。在本文中,我们的目标是最大限度地提高V2I链路的总容量,同时保证了V2V链路的严格传输延迟和可靠性约束。考虑到在高移动式车辆环境中的快速频道变化和连续值状态,我们使用多代理双深度Q学习(DDQN)算法。每个V2V链路被视为代理,通过与环境进行交互,使用更新的Q网络学习最佳策略。实验验证了我们算法的收敛性。仿真结果表明,该方案可以显着优化V2I链路的总容量,并确保V2V链路的延迟和可靠性要求。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号