CASR:优化无服务器计算中的冷启动和资源利用

IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS
Yu Chen , Bo Liu , Weiwei Lin , Yulin Guo , Zhiping Peng
{"title":"CASR:优化无服务器计算中的冷启动和资源利用","authors":"Yu Chen ,&nbsp;Bo Liu ,&nbsp;Weiwei Lin ,&nbsp;Yulin Guo ,&nbsp;Zhiping Peng","doi":"10.1016/j.future.2025.107851","DOIUrl":null,"url":null,"abstract":"<div><div>Serverless computing, also known as Functions as a Service (FaaS), is an emerging cloud deployment paradigm that offers advantages such as pay-as-you-go pricing and automatic scaling. Functions often suffer from cold starts delays due to the overhead of initializing code and data dependencies before execution. Retaining containers in memory for a period after execution can reduce cold start latency. However, existing application-layer solutions overlook the effect of cold start overhead on the availability of containers, resulting in suboptimal balance between cold start latency and memory resource utilization. Furthermore, these strategies typically overlook the optimization of overall cold start overhead, which is essential for enhancing system efficiency. To address these challenges, we propose the Cache-Based Adaptive Scheduler for Serverless Runtime (CASR), an adaptive strategy for managing container runtime configurations. CASR effectively balances cold start latency and memory utilization while reducing overall cold start overhead. Specifically, we introduce a serverless cache (S-Cache) that leverages the equivalence between caching problems and container keep-alive strategies to mitigate cold starts. Additionally, we develop a deep reinforcement learning model, based on the proximal policy optimization algorithm, to enable the automatic scaling of the S-Cache queue, allowing adaptation to dynamic cloud workloads and enhancing memory resource utilization. Extensive simulations on an Azure dataset show that CASR reduces cold starts by 38.75%, improves memory resource utilization by 46.73%, and decreases cold start overhead by 48.53% compared to existing container keep-alive strategies in serverless platforms under common workloads.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"170 ","pages":"Article 107851"},"PeriodicalIF":6.2000,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"CASR: Optimizing cold start and resources utilization in serverless computing\",\"authors\":\"Yu Chen ,&nbsp;Bo Liu ,&nbsp;Weiwei Lin ,&nbsp;Yulin Guo ,&nbsp;Zhiping Peng\",\"doi\":\"10.1016/j.future.2025.107851\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Serverless computing, also known as Functions as a Service (FaaS), is an emerging cloud deployment paradigm that offers advantages such as pay-as-you-go pricing and automatic scaling. Functions often suffer from cold starts delays due to the overhead of initializing code and data dependencies before execution. Retaining containers in memory for a period after execution can reduce cold start latency. However, existing application-layer solutions overlook the effect of cold start overhead on the availability of containers, resulting in suboptimal balance between cold start latency and memory resource utilization. Furthermore, these strategies typically overlook the optimization of overall cold start overhead, which is essential for enhancing system efficiency. To address these challenges, we propose the Cache-Based Adaptive Scheduler for Serverless Runtime (CASR), an adaptive strategy for managing container runtime configurations. CASR effectively balances cold start latency and memory utilization while reducing overall cold start overhead. Specifically, we introduce a serverless cache (S-Cache) that leverages the equivalence between caching problems and container keep-alive strategies to mitigate cold starts. Additionally, we develop a deep reinforcement learning model, based on the proximal policy optimization algorithm, to enable the automatic scaling of the S-Cache queue, allowing adaptation to dynamic cloud workloads and enhancing memory resource utilization. Extensive simulations on an Azure dataset show that CASR reduces cold starts by 38.75%, improves memory resource utilization by 46.73%, and decreases cold start overhead by 48.53% compared to existing container keep-alive strategies in serverless platforms under common workloads.</div></div>\",\"PeriodicalId\":55132,\"journal\":{\"name\":\"Future Generation Computer Systems-The International Journal of Escience\",\"volume\":\"170 \",\"pages\":\"Article 107851\"},\"PeriodicalIF\":6.2000,\"publicationDate\":\"2025-04-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Future Generation Computer Systems-The International Journal of Escience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167739X25001463\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X25001463","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0

摘要

无服务器计算,也称为功能即服务(FaaS),是一种新兴的云部署范例,它提供了即用即付的定价和自动扩展等优势。由于在执行前初始化代码和数据依赖关系的开销,函数经常遭受冷启动延迟。在执行后将容器保留在内存中一段时间可以减少冷启动延迟。但是,现有的应用层解决方案忽略了冷启动开销对容器可用性的影响,导致冷启动延迟和内存资源利用率之间的平衡不是最优的。此外,这些策略通常忽略了整体冷启动开销的优化,而这对于提高系统效率至关重要。为了应对这些挑战,我们提出了基于缓存的无服务器运行时自适应调度器(CASR),这是一种管理容器运行时配置的自适应策略。CASR有效地平衡了冷启动延迟和内存使用,同时降低了总体冷启动开销。具体来说,我们引入了一个无服务器缓存(S-Cache),它利用缓存问题和容器保持活动策略之间的等价性来减轻冷启动。此外,我们开发了一个基于近端策略优化算法的深度强化学习模型,以实现S-Cache队列的自动缩放,从而适应动态云工作负载并提高内存资源利用率。在Azure数据集上进行的大量模拟表明,与无服务器平台中现有的容器保持活动策略相比,CASR减少了38.75%的冷启动,提高了46.73%的内存资源利用率,并减少了48.53%的冷启动开销。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
CASR: Optimizing cold start and resources utilization in serverless computing
Serverless computing, also known as Functions as a Service (FaaS), is an emerging cloud deployment paradigm that offers advantages such as pay-as-you-go pricing and automatic scaling. Functions often suffer from cold starts delays due to the overhead of initializing code and data dependencies before execution. Retaining containers in memory for a period after execution can reduce cold start latency. However, existing application-layer solutions overlook the effect of cold start overhead on the availability of containers, resulting in suboptimal balance between cold start latency and memory resource utilization. Furthermore, these strategies typically overlook the optimization of overall cold start overhead, which is essential for enhancing system efficiency. To address these challenges, we propose the Cache-Based Adaptive Scheduler for Serverless Runtime (CASR), an adaptive strategy for managing container runtime configurations. CASR effectively balances cold start latency and memory utilization while reducing overall cold start overhead. Specifically, we introduce a serverless cache (S-Cache) that leverages the equivalence between caching problems and container keep-alive strategies to mitigate cold starts. Additionally, we develop a deep reinforcement learning model, based on the proximal policy optimization algorithm, to enable the automatic scaling of the S-Cache queue, allowing adaptation to dynamic cloud workloads and enhancing memory resource utilization. Extensive simulations on an Azure dataset show that CASR reduces cold starts by 38.75%, improves memory resource utilization by 46.73%, and decreases cold start overhead by 48.53% compared to existing container keep-alive strategies in serverless platforms under common workloads.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
19.90
自引率
2.70%
发文量
376
审稿时长
10.6 months
期刊介绍: Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications. Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration. Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信