Distributed Reduced Convolution Neural Networks

Mohammad Alajanbi, D. Malerba, He Liu
{"title":"Distributed Reduced Convolution Neural Networks","authors":"Mohammad Alajanbi, D. Malerba, He Liu","doi":"10.58496/mjbd/2021/005","DOIUrl":null,"url":null,"abstract":"The fields of pattern recognition and machine learning frequently make use of something called a Convolution Neural Network, or CNN for short. The kernel extension of CNN, often known as KCNN, offers a performance that is superior to that of conventional CNN. When working with a large-size kernel matrix, the KCNN takes a lot of time and requires a lot of memory, despite the fact that it is capable of solving difficult nonlinear problems. The implementation of a reduced kernel approach has the potential to significantly lower the amount of computational burden and memory consumption. However, since the total quantity of training data continues to expand at an exponential rate, it becomes impossible for a single worker to store the kernel matrix in an efficient manner. This renders centralized data mining impossible to implement. A distributed reduced kernel approach for training CNN on decentralized data, which is referred to as DRCNN, is proposed in this study. In the DRCNN, we will arbitrarily distribute the data to the various nodes. The communication between nodes is static and does not depend on the amount of training data stored on each node; instead, it is determined by the architecture of the network. In contrast to the reduced kernel CNN that is already in use, the DRCNN is a completely distributed training technique that is based on the approach of alternating direction multiplier (ADMM). Experimentation with the large size data set reveals that the distributed technique can produce virtually the same outcomes as the centralized algorithm, and it even requires less time to a significant amount. It results in a significant decrease in the amount of time needed for computation.","PeriodicalId":325612,"journal":{"name":"Mesopotamian Journal of Big Data","volume":"144 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Mesopotamian Journal of Big Data","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.58496/mjbd/2021/005","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7

Abstract

The fields of pattern recognition and machine learning frequently make use of something called a Convolution Neural Network, or CNN for short. The kernel extension of CNN, often known as KCNN, offers a performance that is superior to that of conventional CNN. When working with a large-size kernel matrix, the KCNN takes a lot of time and requires a lot of memory, despite the fact that it is capable of solving difficult nonlinear problems. The implementation of a reduced kernel approach has the potential to significantly lower the amount of computational burden and memory consumption. However, since the total quantity of training data continues to expand at an exponential rate, it becomes impossible for a single worker to store the kernel matrix in an efficient manner. This renders centralized data mining impossible to implement. A distributed reduced kernel approach for training CNN on decentralized data, which is referred to as DRCNN, is proposed in this study. In the DRCNN, we will arbitrarily distribute the data to the various nodes. The communication between nodes is static and does not depend on the amount of training data stored on each node; instead, it is determined by the architecture of the network. In contrast to the reduced kernel CNN that is already in use, the DRCNN is a completely distributed training technique that is based on the approach of alternating direction multiplier (ADMM). Experimentation with the large size data set reveals that the distributed technique can produce virtually the same outcomes as the centralized algorithm, and it even requires less time to a significant amount. It results in a significant decrease in the amount of time needed for computation.
分布式简化卷积神经网络
模式识别和机器学习领域经常使用卷积神经网络,简称CNN。CNN的内核扩展,通常被称为KCNN,提供了优于传统CNN的性能。当处理大尺寸的核矩阵时,KCNN花费了大量的时间和内存,尽管它能够解决困难的非线性问题。精简内核方法的实现有可能显著降低计算负担和内存消耗。然而,由于训练数据的总量继续以指数速率扩展,单个工作人员不可能以有效的方式存储核矩阵。这使得集中式数据挖掘无法实现。本文提出了一种用于在分散数据上训练CNN的分布式简化核方法,称为DRCNN。在DRCNN中,我们将数据任意分布到各个节点。节点之间的通信是静态的,不依赖于每个节点上存储的训练数据量;相反,它是由网络的架构决定的。与已经使用的简化核CNN相比,DRCNN是一种基于交替方向乘法器(ADMM)方法的完全分布式训练技术。对大型数据集的实验表明,分布式技术实际上可以产生与集中式算法相同的结果,而且它甚至需要更少的时间。它可以显著减少计算所需的时间。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信