Stateful Versus Stateless Selection of Edge or Cloud Servers Under Latency Constraints

V. Mancuso, P. Castagno, M. Sereno, M. Marsan
{"title":"Stateful Versus Stateless Selection of Edge or Cloud Servers Under Latency Constraints","authors":"V. Mancuso, P. Castagno, M. Sereno, M. Marsan","doi":"10.1109/WoWMoM54355.2022.00022","DOIUrl":null,"url":null,"abstract":"We consider a radio access network slice serving mobile users whose requests imply computing requirements. Service is virtualized over either a powerful but distant cloud infrastructure or an edge computing host. The latter provides less computing and storage capacity with respect to the cloud, but can be reached with much lower delay. A tradeoff thus naturally arises between computing capacity and data transfer latency. We investigate the performance of this service model, discussing how service requests should be routed to edge or cloud servers. We look at the performance of various classes of online algorithms based on different levels of information about the system state. Our investigation is based on analytical models, simulations in OMNeT++, and a prototype implementation over operational cellular networks. First of all, we observe that distributing the load of service requests over edge and cloud is in general beneficial for performance, and simple to implement with a stateless online server selection policy that can be easily configured with near-optimal performance. Second, we shed light on the limited improvements that stateful polices can offer, notwithstanding they base their decisions on the knowledge of server congestion levels or round-trip latency conditions. Third, we unveil that stateful policies are dangerously prone to errors, which may make stateless policies preferable.","PeriodicalId":275324,"journal":{"name":"2022 IEEE 23rd International Symposium on a World of Wireless, Mobile and Multimedia Networks (WoWMoM)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 23rd International Symposium on a World of Wireless, Mobile and Multimedia Networks (WoWMoM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WoWMoM54355.2022.00022","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

We consider a radio access network slice serving mobile users whose requests imply computing requirements. Service is virtualized over either a powerful but distant cloud infrastructure or an edge computing host. The latter provides less computing and storage capacity with respect to the cloud, but can be reached with much lower delay. A tradeoff thus naturally arises between computing capacity and data transfer latency. We investigate the performance of this service model, discussing how service requests should be routed to edge or cloud servers. We look at the performance of various classes of online algorithms based on different levels of information about the system state. Our investigation is based on analytical models, simulations in OMNeT++, and a prototype implementation over operational cellular networks. First of all, we observe that distributing the load of service requests over edge and cloud is in general beneficial for performance, and simple to implement with a stateless online server selection policy that can be easily configured with near-optimal performance. Second, we shed light on the limited improvements that stateful polices can offer, notwithstanding they base their decisions on the knowledge of server congestion levels or round-trip latency conditions. Third, we unveil that stateful policies are dangerously prone to errors, which may make stateless policies preferable.
延迟约束下边缘或云服务器的有状态与无状态选择
我们考虑一个无线接入网切片服务于移动用户,这些用户的请求包含计算需求。服务可以通过强大但遥远的云基础设施或边缘计算主机进行虚拟化。与云相比,后者提供的计算和存储容量更少,但可以以更低的延迟到达。因此,在计算能力和数据传输延迟之间自然会产生权衡。我们将研究此服务模型的性能,讨论如何将服务请求路由到边缘服务器或云服务器。我们根据不同级别的系统状态信息来研究各种在线算法的性能。我们的研究是基于分析模型,在omnet++中的模拟,以及在可操作的蜂窝网络上的原型实现。首先,我们观察到,在边缘和云上分配服务请求负载通常有利于性能,并且使用无状态在线服务器选择策略很容易实现,可以轻松配置为接近最佳性能。其次,我们阐明了有状态策略所能提供的有限改进,尽管它们的决策是基于服务器拥塞水平或往返延迟条件的知识。第三,我们揭示了有状态策略容易出错的危险,这可能使无状态策略更可取。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信