4.7 Article

AutoMan: Resource-efficient provisioning with tail latency guarantees for microservices

出版社

ELSEVIER
DOI: 10.1016/j.future.2023.01.014

关键词

Cloud computing; Microservices; Resource management; Reinforcement learning; Quality of service; Tail latency

向作者/读者索取更多资源

AutoMan is a learning-driven resource manager for microservices that uses a multi-agent deep deterministic policy gradient (MADDPG) method to efficiently allocate resources while guaranteeing the end-to-end tail latency Service Level Objective (SLO). It proactively identifies critical microservices and performs dynamic reprovisioning to mitigate potential SLO violations. Testbed experiments show that AutoMan can save CPU and memory resources by up to 49.6% and 29.1% on average while ensuring the same end-to-end tail latency objective.
Modern user-facing services are progressively evolving from large monolithic applications to complex graphs of loosely-coupled microservices. While microservice architecture greatly improves the efficiency of development and operation, it also complicates resource allocation and performance guarantee due to complex dependencies across different microservices. To prevent resource wastage and ensure user satisfaction, we present AutoMan, a learning-driven resource manager for microservices that enables much more efficient resource provisioning while guaranteeing the end-to-end tail latency Service Level Objective (SLO). AutoMan leverages a multi-agent deep deterministic policy gradient (MADDPG)-based method to capture the dependencies across different microservices and to allocate a proper amount of resources to each microservice subject to the target end-to-end tail latency SLO. During runtime, it further proactively identifies the critical microservices responsible for performance anomaly by deriving partial SLOs mathematically, and performs dynamic reprovisioning to mitigate the potential SLO violations. Testbed experiments show that AutoMan can save CPU and memory resources by up to 49.6% and 29.1% on average, while guaranteeing the same end-to-end tail latency objective.(c) 2023 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据