首页> 外文期刊>ACM Transactions on Architecture and Code Optimization >Cooperative Multi-Agent Reinforcement Learning-Based Co-optimization of Cores, Caches, and On-chip Network
【24h】

Cooperative Multi-Agent Reinforcement Learning-Based Co-optimization of Cores, Caches, and On-chip Network

机译:基于合作多功能强化学习的核心,高速缓存和片上网络的协同优化

获取原文
获取原文并翻译 | 示例
           

摘要

Modern multi-core systems provide huge computational capabilities, which can be used to run multiple processes concurrently. To achieve the best possible performance within limited power budgets, the various system resources need to be allocated effectively. Any mismatch between runtime resource requirement and allocation leads to a sub-optimal energy-delay product (EDP). Different optimization techniques exist for addressing the problem of mismatch between the dynamic requirement and runtime allocation of the system resources. Choosing between multiple optimizations at runtime is complex due to the non-additive effects, making the scenario suitable for the application of machine learning techniques. We present a novel method, Machine Learned Machines (MLM), by using online reinforcement learning (RL) to perform dynamic partitioning of the last level cache (LLC), along with dynamic voltage and frequency scaling (DVFS) of the core and uncore (interconnection network and LLC). We have proposed and evaluated three different MLM co-optimization techniques based on independent and cooperative multi-agent learners. We show that the co-optimization results in a much lower system EDP than any of the techniques applied individually. We explore various RL models targeted toward optimization of different system metrics and study their effects on a system EDP, system throughput (STP), and Fairness. The various proposed techniques have been extensively evaluated with a mix of 20 workloads on a 4-core system using Spec2006 benchmarks. We have further evaluated our cooperative MLM techniques on a 16-core system. The results show an average of 20.5% and 19.1% system EDP improvement on a 4-core and 16-core system, respectively, with limited degradation of STP and Fairness.
机译:现代多核系统提供巨大的计算能力,可用于同时运行多个进程。为了在有限的电力预算内实现最佳性能,需要有效地分配各种系统资源。运行时资源需求和分配之间的任何不匹配导致子最优能量延迟产品(EDP)。存在不同的优化技术,用于解决系统资源的动态需求和运行时分配之间的不匹配问题。由于非附加效应,在运行时在运行时进行复杂的多个优化选择,使得适用于应用机器学习技术的场景。我们提出了一种新颖的方法,机器学习机(MLM),通过使用在线加强学习(RL)来执行最后一个级别高速缓存(LLC)的动态分区,以及核心和未通知的动态电压和频率缩放(DVF)(互连网络和LLC)。我们提出并评估了基于独立和合作的多助理学习者的三种不同的MLM共同优化技术。我们表明,共同优化导致比单独应用的任何技术更低的系统EDP。我们探索针对不同系统指标优化的各种RL模型,并研究其对系统EDP,系统吞吐量(STP)和公平性的影响。使用SPEC2006基准测试,已经通过20个工作负载的混合进行了广泛评估了各种所提出的技术。我们进一步在16核系统上进行了合作的MLM技术。该结果分别显示平均4核和16核系统的20.5%和19.1%的系统EDP改善,具有限制性的STP和公平性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号