Scalable Edge Computing for Low Latency Data Dissemination in Topic-Based Publish/Subscribe

S. Khare, Hongyang Sun, Kaiwen Zhang, Julien Gascon-Samson, A. Gokhale, X. Koutsoukos, Hamzah Abdelaziz
{"title":"Scalable Edge Computing for Low Latency Data Dissemination in Topic-Based Publish/Subscribe","authors":"S. Khare, Hongyang Sun, Kaiwen Zhang, Julien Gascon-Samson, A. Gokhale, X. Koutsoukos, Hamzah Abdelaziz","doi":"10.1109/SEC.2018.00023","DOIUrl":null,"url":null,"abstract":"Advances in Internet of Things (IoT) give rise to a variety of latency-sensitive, closed-loop applications that reside at the edge. These applications often involve a large number of sensors that generate volumes of data, which must be processed and disseminated in real-time to potentially a large number of entities for actuation, thereby forming a closed-loop, publish-process-subscribe system. To meet the response time requirements of such applications, this paper presents techniques to realize a scalable, fog/edge-based broker architecture that balances data publication and processing loads for topic-based, publish-process-subscribe systems operating at the edge, and assures the Quality-of-Service (QoS), specified as the 90th percentile latency, on a per-topic basis. The key contributions include: (a) a sensitivity analysis to understand the impact of features such as publishing rate, number of subscribers, per-sample processing interval and background load on a topic's performance; (b) a latency prediction model for a set of co-located topics, which is then used for the latency-aware placement of topics on brokers; and (c) an optimization problem formulation for k-topic co-location to minimize the number of brokers while meeting each topic's QoS requirement. Here, k denotes the maximum number of topics that can be placed on a broker. We show that the problem is NP-hard for k >=3 and present three load balancing heuristics. Empirical results are presented to validate the latency prediction model and to evaluate the performance of the proposed heuristics.","PeriodicalId":376439,"journal":{"name":"2018 IEEE/ACM Symposium on Edge Computing (SEC)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"15","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE/ACM Symposium on Edge Computing (SEC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SEC.2018.00023","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 15

Abstract

Advances in Internet of Things (IoT) give rise to a variety of latency-sensitive, closed-loop applications that reside at the edge. These applications often involve a large number of sensors that generate volumes of data, which must be processed and disseminated in real-time to potentially a large number of entities for actuation, thereby forming a closed-loop, publish-process-subscribe system. To meet the response time requirements of such applications, this paper presents techniques to realize a scalable, fog/edge-based broker architecture that balances data publication and processing loads for topic-based, publish-process-subscribe systems operating at the edge, and assures the Quality-of-Service (QoS), specified as the 90th percentile latency, on a per-topic basis. The key contributions include: (a) a sensitivity analysis to understand the impact of features such as publishing rate, number of subscribers, per-sample processing interval and background load on a topic's performance; (b) a latency prediction model for a set of co-located topics, which is then used for the latency-aware placement of topics on brokers; and (c) an optimization problem formulation for k-topic co-location to minimize the number of brokers while meeting each topic's QoS requirement. Here, k denotes the maximum number of topics that can be placed on a broker. We show that the problem is NP-hard for k >=3 and present three load balancing heuristics. Empirical results are presented to validate the latency prediction model and to evaluate the performance of the proposed heuristics.
基于主题的发布/订阅中低延迟数据传播的可扩展边缘计算
物联网(IoT)的进步产生了各种驻留在边缘的对延迟敏感的闭环应用程序。这些应用通常涉及产生大量数据的大量传感器,这些数据必须实时处理并分发给潜在的大量实体以驱动,从而形成一个闭环,发布-流程-订阅系统。为了满足此类应用程序的响应时间需求,本文提出了实现可扩展的基于雾/边缘的代理体系结构的技术,该体系结构为在边缘运行的基于主题的发布-流程-订阅系统平衡数据发布和处理负载,并确保服务质量(QoS),指定为每个主题的第90个百分位延迟。主要贡献包括:(a)敏感性分析,以了解诸如发布率、订阅者数量、每样本处理间隔和背景负载等特征对主题性能的影响;(b)一组共置主题的延迟预测模型,然后用于在代理上对主题的延迟感知放置;(c) k-topic协同定位的优化问题公式,以最小化代理数量,同时满足每个主题的QoS要求。这里,k表示可以放置在代理上的主题的最大数量。我们证明了当k >=3时问题是np困难的,并提出了三种负载平衡启发式方法。实验结果验证了延迟预测模型,并评估了所提出的启发式算法的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信