Crowdlearning: Crowded Deep Learning with Data Privacy

Linlin Chen, Taeho Jung, Haohua Du, Jianwei Qian, Jiahui Hou, Xiangyang Li
{"title":"Crowdlearning: Crowded Deep Learning with Data Privacy","authors":"Linlin Chen, Taeho Jung, Haohua Du, Jianwei Qian, Jiahui Hou, Xiangyang Li","doi":"10.1109/SAHCN.2018.8397100","DOIUrl":null,"url":null,"abstract":"Deep Learning has shown promising performance in a variety of pattern recognition tasks owning to large quantities of training data and complex structures of neural networks. However conventional deep neural network (DNN) training involves centrally collecting and storing the training data, and then centrally training the neural network, which raises much privacy concerns for the data producers. In this paper, we study how to enable deep learning without disclosing individual data to the DNN trainer. We analyze the risks in conventional deep learning training, then propose a novel idea - Crowdlearning, which decentralizes the heavy- load training procedure and deploys the training into a crowd of computation-restricted mobile devices who generate the training data. Finally, we propose SliceNet, which ensures mobile devices can afford the computation cost and simultaneously minimize the total communication cost. The combination of Crowdlearning and SliceNet ensures the sensitive data generated by mobile devices never leave the devices, and the training procedure will hardly disclose any inferable contents. We numerically simulate our prototype of SliceNet which crowdlearns an accurate DNN for image classification, and demonstrate the high performance, acceptable calculation and communication cost, satisfiable privacy protection, and preferable convergence rate, on the benchmark DNN structure and dataset.","PeriodicalId":139623,"journal":{"name":"2018 15th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 15th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SAHCN.2018.8397100","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5

Abstract

Deep Learning has shown promising performance in a variety of pattern recognition tasks owning to large quantities of training data and complex structures of neural networks. However conventional deep neural network (DNN) training involves centrally collecting and storing the training data, and then centrally training the neural network, which raises much privacy concerns for the data producers. In this paper, we study how to enable deep learning without disclosing individual data to the DNN trainer. We analyze the risks in conventional deep learning training, then propose a novel idea - Crowdlearning, which decentralizes the heavy- load training procedure and deploys the training into a crowd of computation-restricted mobile devices who generate the training data. Finally, we propose SliceNet, which ensures mobile devices can afford the computation cost and simultaneously minimize the total communication cost. The combination of Crowdlearning and SliceNet ensures the sensitive data generated by mobile devices never leave the devices, and the training procedure will hardly disclose any inferable contents. We numerically simulate our prototype of SliceNet which crowdlearns an accurate DNN for image classification, and demonstrate the high performance, acceptable calculation and communication cost, satisfiable privacy protection, and preferable convergence rate, on the benchmark DNN structure and dataset.
众学习:具有数据隐私的拥挤深度学习
由于大量的训练数据和复杂的神经网络结构,深度学习在各种模式识别任务中表现出了良好的性能。然而,传统的深度神经网络(DNN)训练涉及集中收集和存储训练数据,然后集中训练神经网络,这给数据生产者带来了很多隐私问题。在本文中,我们研究了如何在不向DNN训练器透露个人数据的情况下实现深度学习。在分析传统深度学习训练存在的风险的基础上,提出了一种新的思想——众学习(Crowdlearning),该思想将繁重的训练过程分散开来,并将训练部署到一群计算受限的移动设备上,由这些设备生成训练数据。最后,我们提出了SliceNet,以确保移动设备能够承担计算成本,同时最小化总通信成本。Crowdlearning与SliceNet的结合,保证了移动设备产生的敏感数据永远不会离开设备,训练过程中几乎不会泄露任何可推断的内容。通过对SliceNet原型进行数值模拟,验证了该模型在基准DNN结构和数据集上的高性能、可接受的计算和通信成本、令人满意的隐私保护和较好的收敛速度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信