Distributed Deep Learning on Wimpy Smartphone Nodes

Tzoof Hemed, Nitai Lavie, R. Kaplan
{"title":"Distributed Deep Learning on Wimpy Smartphone Nodes","authors":"Tzoof Hemed, Nitai Lavie, R. Kaplan","doi":"10.1109/ICSEE.2018.8646195","DOIUrl":null,"url":null,"abstract":"Deep Neural Networks (DNN), contain multiple convolutional and several fully connected layers, require considerable hardware resources to train in a reasonable time. Multiple CPUs, GPUs or FPGAs are usually combined to reduce the training time of a DNN. However, many individuals or small organizations do not possess the resources to obtain multiple hardware units.The contribution of this work is two-fold. First, we present an implementation of a distributed DNN training system that uses multiple small (wimpy) nodes to accelerate the training process. The nodes are mobile smartphone devices, with variable hardware specifications. All DNN training tasks are performed on the small nodes, coordinated by a centralized server. Second, we propose a novel method to mitigate issues arising from the variability in hardware resources. We demonstrate that the method allows training a DNN to high accuracy on known image recognition datasets with multiple small different nodes. The proposed method factors in the contribution from each node according to its run time on a specific training task, relative to the other nodes. In addition, we discuss practical challenges that arise from small node system and suggest several solutions.","PeriodicalId":254455,"journal":{"name":"2018 IEEE International Conference on the Science of Electrical Engineering in Israel (ICSEE)","volume":"409 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE International Conference on the Science of Electrical Engineering in Israel (ICSEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSEE.2018.8646195","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Deep Neural Networks (DNN), contain multiple convolutional and several fully connected layers, require considerable hardware resources to train in a reasonable time. Multiple CPUs, GPUs or FPGAs are usually combined to reduce the training time of a DNN. However, many individuals or small organizations do not possess the resources to obtain multiple hardware units.The contribution of this work is two-fold. First, we present an implementation of a distributed DNN training system that uses multiple small (wimpy) nodes to accelerate the training process. The nodes are mobile smartphone devices, with variable hardware specifications. All DNN training tasks are performed on the small nodes, coordinated by a centralized server. Second, we propose a novel method to mitigate issues arising from the variability in hardware resources. We demonstrate that the method allows training a DNN to high accuracy on known image recognition datasets with multiple small different nodes. The proposed method factors in the contribution from each node according to its run time on a specific training task, relative to the other nodes. In addition, we discuss practical challenges that arise from small node system and suggest several solutions.
弱智能手机节点上的分布式深度学习
深度神经网络(DNN)包含多个卷积层和多个全连接层,需要大量的硬件资源才能在合理的时间内进行训练。为了减少深度神经网络的训练时间,通常将多个cpu、gpu或fpga组合在一起。然而,许多个人或小型组织并不拥有获得多个硬件单元的资源。这项工作的贡献是双重的。首先,我们提出了一个分布式DNN训练系统的实现,该系统使用多个小(弱)节点来加速训练过程。节点是具有可变硬件规格的移动智能手机设备。所有DNN训练任务都在小节点上执行,由中央服务器协调。其次,我们提出了一种新的方法来缓解硬件资源的可变性所带来的问题。我们证明了该方法可以在具有多个小不同节点的已知图像识别数据集上训练DNN以达到高精度。所提出的方法根据每个节点在特定训练任务上的运行时间来考虑其相对于其他节点的贡献。此外,我们还讨论了小节点系统带来的实际挑战,并提出了一些解决方案。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信