On the Computational Power of Convolution Pooling: A Theoretical Approach for Deep Learning

K. Nakano, Shotaro Aoki, Yasuaki Ito, Akihiko Kasagi
{"title":"On the Computational Power of Convolution Pooling: A Theoretical Approach for Deep Learning","authors":"K. Nakano, Shotaro Aoki, Yasuaki Ito, Akihiko Kasagi","doi":"10.1109/IPDPSW52791.2021.00100","DOIUrl":null,"url":null,"abstract":"Convolutional neural networks (CNNs) have been widely used for image analysis and recognition. For example, LeNet-5 is a 7-layer convectional neural network, which can attain more than 99% test accuracy for classification of handwritten digits. CNNs repeats convolution and pooling operations alternately. However, the computational capability of such operations is not clear. We are curious to know a class of problems that can be solved by CNNs. As a formal approach for this task, we introduce a theoretical parallel computational model of CNNs that we call the convolution-pooling machine. It captures the essence of convolution and pooling operations, and application of non-linear activation functions performed in CNNs. In this paper, we assume the convolution-pooling machine operating on 1-dimensional arrays for simplicity, and focus on the problem of classification of inputs by the distance of two feature points. More specifically, we will design a convolution-pooling machine solving the problem Dk (k≥1), a problem to determine if the distance of the two 1’s is at most k or not. For designing the convolution-pooling machine solving the problem Dk, we generate a mixed-integer linear programming problem (MILP) with constraints and objective functions. We have solved the generated linear programming problem for each Dk (1≤k≤128) by Gurobi optimizer, a commercial MILP solver. We succeeded in finding a solution for all Dk (1 ≤ k ≤ 128) and designing the convolution-pooling machine for solving them. This fact indicates that convolution and pooling operations in CNNs may have the computational capability of classification by the distance of feature points.","PeriodicalId":170832,"journal":{"name":"2021 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW)","volume":"82 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IPDPSW52791.2021.00100","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Convolutional neural networks (CNNs) have been widely used for image analysis and recognition. For example, LeNet-5 is a 7-layer convectional neural network, which can attain more than 99% test accuracy for classification of handwritten digits. CNNs repeats convolution and pooling operations alternately. However, the computational capability of such operations is not clear. We are curious to know a class of problems that can be solved by CNNs. As a formal approach for this task, we introduce a theoretical parallel computational model of CNNs that we call the convolution-pooling machine. It captures the essence of convolution and pooling operations, and application of non-linear activation functions performed in CNNs. In this paper, we assume the convolution-pooling machine operating on 1-dimensional arrays for simplicity, and focus on the problem of classification of inputs by the distance of two feature points. More specifically, we will design a convolution-pooling machine solving the problem Dk (k≥1), a problem to determine if the distance of the two 1’s is at most k or not. For designing the convolution-pooling machine solving the problem Dk, we generate a mixed-integer linear programming problem (MILP) with constraints and objective functions. We have solved the generated linear programming problem for each Dk (1≤k≤128) by Gurobi optimizer, a commercial MILP solver. We succeeded in finding a solution for all Dk (1 ≤ k ≤ 128) and designing the convolution-pooling machine for solving them. This fact indicates that convolution and pooling operations in CNNs may have the computational capability of classification by the distance of feature points.
论卷积池的计算能力:一种深度学习的理论方法
卷积神经网络(cnn)在图像分析和识别中得到了广泛的应用。例如,LeNet-5是一个7层的传统神经网络,它对手写数字的分类可以达到99%以上的测试准确率。cnn交替重复卷积和池化操作。然而,这种操作的计算能力尚不清楚。我们很想知道cnn可以解决的一类问题。作为这项任务的正式方法,我们引入了一个cnn的理论并行计算模型,我们称之为卷积池机。它抓住了卷积和池化操作的本质,以及非线性激活函数在cnn中的应用。为了简单起见,本文假设卷积池机在一维数组上运行,并重点研究根据两个特征点的距离对输入进行分类的问题。更具体地说,我们将设计一个卷积池机来解决问题Dk (k≥1),这个问题是确定两个1的距离是否最大为k。为了设计解决Dk问题的卷积池机,我们生成了一个带约束和目标函数的混合整数线性规划问题(MILP)。利用商业MILP求解器Gurobi优化器求解了每个Dk(1≤k≤128)的生成线性规划问题。我们成功地找到了所有Dk(1≤k≤128)的解,并设计了求解它们的卷积池机。这表明cnn中的卷积和池化操作可能具有根据特征点距离进行分类的计算能力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信