{"title":"Management of autoscaling serverless functions in edge computing via Q-Learning","authors":"Priscilla Benedetti , Mauro Femminella , Gianluca Reali","doi":"10.1016/j.future.2025.108112","DOIUrl":null,"url":null,"abstract":"<div><div>Serverless computing is a recently introduced deployment model to provide cloud services. The autoscaling of function instances allows adapting allocated resources to workload, so as to reduce latency and improve resource usage efficiency. However, autoscaling mechanisms could be affected by undesired ‘cold starts’ events, causing latency peaks due to spawning of new instances, which can be critical in edge deployments where applications are typically sensitive to latency. In order to regulate autoscaling of functions and mitigate the latency for accessing services, which may hinder the adoption of the serverless model in edge computing, we resort to the usage of reinforcement learning. Our experimental system is based on OpenFaaS, the most popular open-source Kubernetes-based serverless platform. In this system, we introduce a Q-Learning (QL) agent to dynamically configure the Kubernetes Horizontal Pod Autoscaler (HPA). This is accomplished via a QL model state space and a reward function definition that enforce service level agreement (SLA) compliance, in terms of latency, without allocating excessive resources. The agent is trained and tested using real serverless function invocation patterns, made available by Microsoft Azure. The experimental results show the benefits provided by the proposed solution over state-of-the-art in terms of compliance to the SLA, while limiting resource consumption and service request losses.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"175 ","pages":"Article 108112"},"PeriodicalIF":6.2000,"publicationDate":"2025-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X25004066","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
Serverless computing is a recently introduced deployment model to provide cloud services. The autoscaling of function instances allows adapting allocated resources to workload, so as to reduce latency and improve resource usage efficiency. However, autoscaling mechanisms could be affected by undesired ‘cold starts’ events, causing latency peaks due to spawning of new instances, which can be critical in edge deployments where applications are typically sensitive to latency. In order to regulate autoscaling of functions and mitigate the latency for accessing services, which may hinder the adoption of the serverless model in edge computing, we resort to the usage of reinforcement learning. Our experimental system is based on OpenFaaS, the most popular open-source Kubernetes-based serverless platform. In this system, we introduce a Q-Learning (QL) agent to dynamically configure the Kubernetes Horizontal Pod Autoscaler (HPA). This is accomplished via a QL model state space and a reward function definition that enforce service level agreement (SLA) compliance, in terms of latency, without allocating excessive resources. The agent is trained and tested using real serverless function invocation patterns, made available by Microsoft Azure. The experimental results show the benefits provided by the proposed solution over state-of-the-art in terms of compliance to the SLA, while limiting resource consumption and service request losses.
期刊介绍:
Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications.
Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration.
Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.