Exploring compression and parallelization techniques for distribution of deep neural networks over Edge-Fog continuum - a review

Azra Nazir, R. N. Mir, Shaima Qureshi
{"title":"Exploring compression and parallelization techniques for distribution of deep neural networks over Edge-Fog continuum - a review","authors":"Azra Nazir, R. N. Mir, Shaima Qureshi","doi":"10.1108/ijicc-04-2020-0038","DOIUrl":null,"url":null,"abstract":"PurposeThe trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant. But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015. Thus, it is high time to explore a different approach to tackle this issue, keeping in view the characteristics and needs of the two fields. Processing at the Edge can boost applications with real-time deadlines while complementing security.Design/methodology/approachThis review paper contributes towards three cardinal directions of research in the field of DL for IoT. The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices, forming the realm of the things for IoT. The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques. An appropriate combination of these techniques, including regularization, quantization, and pruning, can aid in building an effective compression pipeline for establishing DL models for IoT use-cases. The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.FindingsDL models are growing deeper with every passing year. Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm. It is realized that a vertically partitioned compressed deep model can handle the trade-off between size, accuracy, communication overhead, bandwidth utilization, and latency but at the expense of an additionally considerable memory footprint. To reduce the memory budget, we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks. However, the critical point between accuracy and size for such models needs further investigation.Originality/valueTo the best of our knowledge, no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum. Besides covering techniques and frameworks that have tried to bring inference to the Edge, the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT. The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.","PeriodicalId":352072,"journal":{"name":"Int. J. Intell. Comput. Cybern.","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Int. J. Intell. Comput. Cybern.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1108/ijicc-04-2020-0038","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8

Abstract

PurposeThe trend of “Deep Learning for Internet of Things (IoT)” has gained fresh momentum with enormous upcoming applications employing these models as their processing engine and Cloud as their resource giant. But this picture leads to underutilization of ever-increasing device pool of IoT that has already passed 15 billion mark in 2015. Thus, it is high time to explore a different approach to tackle this issue, keeping in view the characteristics and needs of the two fields. Processing at the Edge can boost applications with real-time deadlines while complementing security.Design/methodology/approachThis review paper contributes towards three cardinal directions of research in the field of DL for IoT. The first section covers the categories of IoT devices and how Fog can aid in overcoming the underutilization of millions of devices, forming the realm of the things for IoT. The second direction handles the issue of immense computational requirements of DL models by uncovering specific compression techniques. An appropriate combination of these techniques, including regularization, quantization, and pruning, can aid in building an effective compression pipeline for establishing DL models for IoT use-cases. The third direction incorporates both these views and introduces a novel approach of parallelization for setting up a distributed systems view of DL for IoT.FindingsDL models are growing deeper with every passing year. Well-coordinated distributed execution of such models using Fog displays a promising future for the IoT application realm. It is realized that a vertically partitioned compressed deep model can handle the trade-off between size, accuracy, communication overhead, bandwidth utilization, and latency but at the expense of an additionally considerable memory footprint. To reduce the memory budget, we propose to exploit Hashed Nets as potentially favorable candidates for distributed frameworks. However, the critical point between accuracy and size for such models needs further investigation.Originality/valueTo the best of our knowledge, no study has explored the inherent parallelism in deep neural network architectures for their efficient distribution over the Edge-Fog continuum. Besides covering techniques and frameworks that have tried to bring inference to the Edge, the review uncovers significant issues and possible future directions for endorsing deep models as processing engines for real-time IoT. The study is directed to both researchers and industrialists to take on various applications to the Edge for better user experience.
探讨边缘雾连续体上深度神经网络分布的压缩和并行化技术
随着大量即将到来的应用将这些模型作为其处理引擎和云作为其资源巨头,“物联网深度学习”的趋势获得了新的动力。但这种情况导致物联网设备池的利用率不断增加,2015年已经超过150亿大关。因此,在考虑到这两个领域的特点和需要的情况下,探索解决这一问题的不同途径是当务之急。Edge上的处理可以提高应用程序的实时截止日期,同时补充安全性。设计/方法/方法这篇综述论文对物联网深度学习领域的三个主要研究方向做出了贡献。第一部分涵盖了物联网设备的类别,以及雾如何帮助克服数百万设备的未充分利用,形成物联网的领域。第二个方向通过揭示特定的压缩技术来处理深度学习模型的巨大计算需求问题。这些技术的适当组合,包括正则化、量化和修剪,可以帮助构建一个有效的压缩管道,为物联网用例建立深度学习模型。第三个方向结合了这两种观点,并引入了一种新的并行化方法,用于为物联网建立分布式系统的深度学习视图。随着时间的推移,sdl模型变得越来越深入。使用Fog对这些模型进行良好协调的分布式执行,为物联网应用领域展示了一个充满希望的未来。人们认识到,垂直分区的压缩深度模型可以处理大小、准确性、通信开销、带宽利用率和延迟之间的权衡,但代价是额外的相当大的内存占用。为了减少内存预算,我们建议利用哈希网络作为分布式框架的潜在有利候选。然而,这种模型的精度和尺寸之间的临界点需要进一步研究。原创性/价值据我们所知,还没有研究探索深度神经网络架构在边缘-雾连续体上有效分布的内在并行性。除了涵盖试图将推理带到边缘的技术和框架外,该评论还揭示了支持深度模型作为实时物联网处理引擎的重要问题和可能的未来方向。这项研究针对的是研究人员和实业家,他们将各种应用程序带到Edge上,以获得更好的用户体验。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信