Yu Chen , Bo Liu , Weiwei Lin , Yulin Guo , Zhiping Peng
{"title":"CASR:优化无服务器计算中的冷启动和资源利用","authors":"Yu Chen , Bo Liu , Weiwei Lin , Yulin Guo , Zhiping Peng","doi":"10.1016/j.future.2025.107851","DOIUrl":null,"url":null,"abstract":"<div><div>Serverless computing, also known as Functions as a Service (FaaS), is an emerging cloud deployment paradigm that offers advantages such as pay-as-you-go pricing and automatic scaling. Functions often suffer from cold starts delays due to the overhead of initializing code and data dependencies before execution. Retaining containers in memory for a period after execution can reduce cold start latency. However, existing application-layer solutions overlook the effect of cold start overhead on the availability of containers, resulting in suboptimal balance between cold start latency and memory resource utilization. Furthermore, these strategies typically overlook the optimization of overall cold start overhead, which is essential for enhancing system efficiency. To address these challenges, we propose the Cache-Based Adaptive Scheduler for Serverless Runtime (CASR), an adaptive strategy for managing container runtime configurations. CASR effectively balances cold start latency and memory utilization while reducing overall cold start overhead. Specifically, we introduce a serverless cache (S-Cache) that leverages the equivalence between caching problems and container keep-alive strategies to mitigate cold starts. Additionally, we develop a deep reinforcement learning model, based on the proximal policy optimization algorithm, to enable the automatic scaling of the S-Cache queue, allowing adaptation to dynamic cloud workloads and enhancing memory resource utilization. Extensive simulations on an Azure dataset show that CASR reduces cold starts by 38.75%, improves memory resource utilization by 46.73%, and decreases cold start overhead by 48.53% compared to existing container keep-alive strategies in serverless platforms under common workloads.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"170 ","pages":"Article 107851"},"PeriodicalIF":6.2000,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"CASR: Optimizing cold start and resources utilization in serverless computing\",\"authors\":\"Yu Chen , Bo Liu , Weiwei Lin , Yulin Guo , Zhiping Peng\",\"doi\":\"10.1016/j.future.2025.107851\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Serverless computing, also known as Functions as a Service (FaaS), is an emerging cloud deployment paradigm that offers advantages such as pay-as-you-go pricing and automatic scaling. Functions often suffer from cold starts delays due to the overhead of initializing code and data dependencies before execution. Retaining containers in memory for a period after execution can reduce cold start latency. However, existing application-layer solutions overlook the effect of cold start overhead on the availability of containers, resulting in suboptimal balance between cold start latency and memory resource utilization. Furthermore, these strategies typically overlook the optimization of overall cold start overhead, which is essential for enhancing system efficiency. To address these challenges, we propose the Cache-Based Adaptive Scheduler for Serverless Runtime (CASR), an adaptive strategy for managing container runtime configurations. CASR effectively balances cold start latency and memory utilization while reducing overall cold start overhead. Specifically, we introduce a serverless cache (S-Cache) that leverages the equivalence between caching problems and container keep-alive strategies to mitigate cold starts. Additionally, we develop a deep reinforcement learning model, based on the proximal policy optimization algorithm, to enable the automatic scaling of the S-Cache queue, allowing adaptation to dynamic cloud workloads and enhancing memory resource utilization. Extensive simulations on an Azure dataset show that CASR reduces cold starts by 38.75%, improves memory resource utilization by 46.73%, and decreases cold start overhead by 48.53% compared to existing container keep-alive strategies in serverless platforms under common workloads.</div></div>\",\"PeriodicalId\":55132,\"journal\":{\"name\":\"Future Generation Computer Systems-The International Journal of Escience\",\"volume\":\"170 \",\"pages\":\"Article 107851\"},\"PeriodicalIF\":6.2000,\"publicationDate\":\"2025-04-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Future Generation Computer Systems-The International Journal of Escience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167739X25001463\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X25001463","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
CASR: Optimizing cold start and resources utilization in serverless computing
Serverless computing, also known as Functions as a Service (FaaS), is an emerging cloud deployment paradigm that offers advantages such as pay-as-you-go pricing and automatic scaling. Functions often suffer from cold starts delays due to the overhead of initializing code and data dependencies before execution. Retaining containers in memory for a period after execution can reduce cold start latency. However, existing application-layer solutions overlook the effect of cold start overhead on the availability of containers, resulting in suboptimal balance between cold start latency and memory resource utilization. Furthermore, these strategies typically overlook the optimization of overall cold start overhead, which is essential for enhancing system efficiency. To address these challenges, we propose the Cache-Based Adaptive Scheduler for Serverless Runtime (CASR), an adaptive strategy for managing container runtime configurations. CASR effectively balances cold start latency and memory utilization while reducing overall cold start overhead. Specifically, we introduce a serverless cache (S-Cache) that leverages the equivalence between caching problems and container keep-alive strategies to mitigate cold starts. Additionally, we develop a deep reinforcement learning model, based on the proximal policy optimization algorithm, to enable the automatic scaling of the S-Cache queue, allowing adaptation to dynamic cloud workloads and enhancing memory resource utilization. Extensive simulations on an Azure dataset show that CASR reduces cold starts by 38.75%, improves memory resource utilization by 46.73%, and decreases cold start overhead by 48.53% compared to existing container keep-alive strategies in serverless platforms under common workloads.
期刊介绍:
Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications.
Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration.
Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.