首页> 外文期刊>Wireless Networks >Deep reinforcement learning-based computation offloading and resource allocation in security-aware mobile edge computing
【24h】

Deep reinforcement learning-based computation offloading and resource allocation in security-aware mobile edge computing

机译:基于深度加强学习的基于学习的计算卸载和安全感知移动边缘计算中的资源分配

获取原文
获取原文并翻译 | 示例
           

摘要

Owing to the insufficient processing ability of wireless devices (WDs), it is difficult for WDs to process these data within the deadline associated with the quality of service requirements. Offloading computation tasks (workloads) to emerging mobile edge computing servers with small or macro base stations is an effective and feasible solution. However, the offloaded data will be fully exposed and vulnerable to security threats. In this paper, we introduce a wireless communication and computation model of partial computation offloading and resource allocation considering the time-varying channel state, the bandwidth constraint, the stochastic arrival of workloads, and privacy preservation. To simultaneously optimize the computation and execution delays, the power consumption, and the bandwidth resources, we model the optimization problem as a Markov decision process (MDP) to minimize the weighted sum cost of the system. Owing to the difficult problems of lack of priori knowledge and the curse of dimensionality, we propose a decentralized optimization scheme on partial computation offloading and resource allocation based on deep reinforcement learning (DOCRRL). According to the time-varying channel state, the arrival rate of computation workloads, and the signal-to-interference-plus-noise ratio, the DOCRRL algorithm can learn the optimal policy for decision-making under stringent latency and risk constraints that prevent the curse of dimensionality from arising owing to the high-dimensional action space and state space. The numerical results reveal that DOCRRL can explore and learn the optimal decision-making policy without priori knowledge; it outperforms four baseline schemes in simulation environments.
机译:由于无线设备的处理能力不足(WDS),WDS难以在与服务质量要求相关联的截止日期内处理这些数据。将计算任务(工作负载)卸载到具有小型或宏基站的新兴移动边缘计算服务器是一种有效且可行的解决方案。但是,卸载数据将完全暴露并容易受到安全威胁的影响。在本文中,考虑时变信道状态,带宽约束,工作负载随机到达以及隐私保存,引入部分计算卸载和资源分配的无线通信和计算模型。为了同时优化计算和执行延迟,功耗和带宽资源,我们将优化问题模拟为Markov决策过程(MDP),以最小化系统的加权成本。由于缺乏先验知识的问题和维度的诅咒,我们提出了一种基于深度加强学习(DOCRRL)的部分计算卸载和资源分配的分散优化方案。根据时变信道状态,计算工作负载的到达率,以及信号到干扰 - 加噪声比,DoCRRL算法可以在严格的延迟和风险限制下学习决策的最佳策略,以防止由于高维行动空间和状态空间而产生的维度诅咒。数值结果表明,Docrrl可以探索并学习无需先验知识的最佳决策政策;它在仿真环境中优于四种基线方案。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号