Reinforcement Learning Based Vertical Scaling for Hybrid Deployment in Cloud Computing

Published: 01 Jan 2022, Last Modified: 13 Nov 2024BIC-TA 2022EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: To maximize the CPU utilization of the server, offline tasks are usually deployed to the same server where the online service is running. Considering the necessity to ensure the service quality of online services, it is common practice to isolate the resources of online services. How to set the resource quota for online services not only affects the service quality of online services, but also affects the number and the stability of offline tasks that can be run on the server. Traditional rule-based methods or prediction-based methods will cause over-provision and fail to consider the stability of offline tasks, which often cannot achieve stability and efficiency. In this paper, reinforcement learning is proposed for the first time to solve the hybrid deployment of online services and offline tasks and dynamically adjust the resource quota of online services more effectively. Compared with the original state of the server, our proposed method reduces CPU idleness rate by 35.32% and increases CPU resource utilization rate by 3.84%.
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview