{"title":"Distributed Reduced Convolution Neural Networks","authors":"Mohammad Alajanbi, D. Malerba, He Liu","doi":"10.58496/mjbd/2021/005","DOIUrl":null,"url":null,"abstract":"The fields of pattern recognition and machine learning frequently make use of something called a Convolution Neural Network, or CNN for short. The kernel extension of CNN, often known as KCNN, offers a performance that is superior to that of conventional CNN. When working with a large-size kernel matrix, the KCNN takes a lot of time and requires a lot of memory, despite the fact that it is capable of solving difficult nonlinear problems. The implementation of a reduced kernel approach has the potential to significantly lower the amount of computational burden and memory consumption. However, since the total quantity of training data continues to expand at an exponential rate, it becomes impossible for a single worker to store the kernel matrix in an efficient manner. This renders centralized data mining impossible to implement. A distributed reduced kernel approach for training CNN on decentralized data, which is referred to as DRCNN, is proposed in this study. In the DRCNN, we will arbitrarily distribute the data to the various nodes. The communication between nodes is static and does not depend on the amount of training data stored on each node; instead, it is determined by the architecture of the network. In contrast to the reduced kernel CNN that is already in use, the DRCNN is a completely distributed training technique that is based on the approach of alternating direction multiplier (ADMM). Experimentation with the large size data set reveals that the distributed technique can produce virtually the same outcomes as the centralized algorithm, and it even requires less time to a significant amount. It results in a significant decrease in the amount of time needed for computation.","PeriodicalId":325612,"journal":{"name":"Mesopotamian Journal of Big Data","volume":"144 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Mesopotamian Journal of Big Data","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.58496/mjbd/2021/005","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7
Abstract
The fields of pattern recognition and machine learning frequently make use of something called a Convolution Neural Network, or CNN for short. The kernel extension of CNN, often known as KCNN, offers a performance that is superior to that of conventional CNN. When working with a large-size kernel matrix, the KCNN takes a lot of time and requires a lot of memory, despite the fact that it is capable of solving difficult nonlinear problems. The implementation of a reduced kernel approach has the potential to significantly lower the amount of computational burden and memory consumption. However, since the total quantity of training data continues to expand at an exponential rate, it becomes impossible for a single worker to store the kernel matrix in an efficient manner. This renders centralized data mining impossible to implement. A distributed reduced kernel approach for training CNN on decentralized data, which is referred to as DRCNN, is proposed in this study. In the DRCNN, we will arbitrarily distribute the data to the various nodes. The communication between nodes is static and does not depend on the amount of training data stored on each node; instead, it is determined by the architecture of the network. In contrast to the reduced kernel CNN that is already in use, the DRCNN is a completely distributed training technique that is based on the approach of alternating direction multiplier (ADMM). Experimentation with the large size data set reveals that the distributed technique can produce virtually the same outcomes as the centralized algorithm, and it even requires less time to a significant amount. It results in a significant decrease in the amount of time needed for computation.