首页> 外文会议>IEEE international conference on distributed computing systemss >Interference-Aware Component Scheduling for Reducing Tail Latency in Cloud Interactive Services
【24h】

Interference-Aware Component Scheduling for Reducing Tail Latency in Cloud Interactive Services

机译:减少云交互服务中的尾部延迟的可识别干扰的组件计划

获取原文

摘要

Large-scale interactive services usually divide requests into multiple sub-requests and distribute them to a large number of server components for parallel execution. Hence the tail latency (i.e. The slowest component's latency) of these components determines the overall service latency. On a cloud platform, each component shares and competes node resources such as caches and I/O bandwidths with its co-located jobs, hence inevitably suffering from their performance interference. In this paper, we study the short-running jobs in a 12k-node Google cluster to illustrate the dynamic resource demands of these jobs, resulting in both individual components' latency variability over time and across different nodes and hence posing a major challenge to maintain low tail latency. Given this motivation, this paper introduces a dynamic and interference-aware scheduler for large-scale, parallel cloud services. At each scheduling interval, it collects workload and resource contention information of a running service, and predicts both the component latency on different nodes and the overall service performance. Based on the predicted performance, the scheduler identifies straggling components and conducts near-optimal component-node allocations to adapt to the changing workloads and performance interferences. We demonstrate that, using realistic workloads, the proposed approach achieves significant reductions in tail latency compared to the basic approach without scheduling.
机译:大型交互式服务通常将请求分为多个子请求,然后将它们分发到大量服务器组件以并行执行。因此,这些组件的尾部等待时间(即最慢组件的等待时间)决定了整体服务等待时间。在云平台上,每个组件都与其并置的作业共享并竞争节点资源(例如缓存和I / O带宽),因此不可避免地会遭受其性能干扰。在本文中,我们研究了一个12k节点Google集群中的短期作业,以说明这些作业的动态资源需求,从而导致各个组件的延迟随时间推移以及跨不同节点的变化,因此在维护方面构成了重大挑战低尾部等待时间。鉴于这种动机,本文介绍了一种用于大型并行云服务的动态且可感知干扰的调度程序。在每个调度间隔,它都会收集正在运行的服务的工作负载和资源争用信息,并预测不同节点上的组件延迟以及整体服务性能。基于预测的性能,调度程序可以识别散乱的组件,并进行接近最佳的组件节点分配,以适应不断变化的工作负载和性能干扰。我们证明,与没有调度的基本方法相比,使用现实的工作量,所提出的方法可显着减少尾部等待时间。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号