当前位置: X-MOL 学术IEEE Netw. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Going Fast and Fair: Latency Optimization for Cloud-Based Service Chains
IEEE NETWORK ( IF 9.3 ) Pub Date : 2017-11-29 , DOI: 10.1109/mnet.2017.1700275
Yuchao Zhang , Ke Xu , Haiyang Wang , Qi Li , Tong Li , Xuan Cao

State-of-the-art microservices have been attracting more attention in recent years. A broad spectrum of online interactive applications are now programmed to service chains on the cloud, seeking better system scalability and lower operating costs. Different from the conventional batch jobs, most of these applications consist of multiple stand-alone services that communicate with each other. These step-by-step operations unavoidably introduce higher latency to the delay-sensitive chained services. In this article, we aim at designing an optimization approach for reducing the latency of chained services. Specifically, presenting the measurement and analysis of chained services on Baidu's cloud platform, our real-world trace indicates that these chained services are suffering from significantly high latency because they are mostly handled by different queues on cloud servers for multiple times. However, such a unique feature introduces significant challenges to optimize a microservice's overall queueing delay. To address this problem, we propose a delay-guaranteed approach to accelerate the overall queueing of chained services while obtaining fairness across all the workloads. Our evaluations on Baidu servers shows that the proposed design can successfully reduce the latency of chained services by 35 percent with minimal impact on other workloads.

中文翻译:

快速,公平地进行:基于云的服务链的延迟优化

近年来,最先进的微服务引起了更多关注。现在,广泛的在线交互式应用程序已被编程为云上的服务链,以寻求更好的系统可扩展性和更低的运营成本。与传统的批处理作业不同,这些应用程序中的大多数由相互通信的多个独立服务组成。这些分步操作不可避免地会给延迟敏感的链接服务带来更高的延迟。在本文中,我们旨在设计一种优化方法来减少链接服务的延迟。具体来说,在百度的云平台上介绍链式服务的度量和分析,我们的真实记录表明,这些链接服务正遭受很高的延迟,因为它们大多是由云服务器上的不同队列多次处理的。但是,这种独特的功能给优化微服务的整体排队延迟带来了重大挑战。为了解决此问题,我们提出了一种延迟保证的方法,以加快链式服务的整体排队,同时在所有工作负载之间获得公平性。我们对百度服务器的评估表明,所提出的设计可以成功地将链接服务的延迟减少35%,而对其他工作负载的影响却最小。我们提出了一种延迟保证的方法,以加快链式服务的整体排队,同时在所有工作负载之间获得公平。我们对百度服务器的评估表明,所提出的设计可以成功地将链接服务的延迟减少35%,而对其他工作负载的影响却最小。我们提出了一种延迟保证的方法,以加快链式服务的整体排队,同时在所有工作负载之间获得公平。我们对百度服务器的评估表明,所提出的设计可以成功地将链接服务的延迟减少35%,而对其他工作负载的影响却最小。
更新日期:2018-04-03
down
wechat
bug