Exploring Unexplored Tensor Network Decompositions for Convolutional Neural Networks

K. Hayashi, Taiki Yamaguchi, Yohei Sugawara, S. Maeda
{"title":"Exploring Unexplored Tensor Network Decompositions for Convolutional Neural Networks","authors":"K. Hayashi, Taiki Yamaguchi, Yohei Sugawara, S. Maeda","doi":"10.3902/jnns.29.193","DOIUrl":null,"url":null,"abstract":"Tensor decomposition methods are widely used for model compression and fast inference in convolutional neural networks (CNNs). Although many decompositions are conceivable, only CP decomposition and a few others have been applied in practice, and no extensive comparisons have been made between available methods. Previous studies have not determined how many decompositions are available, nor which of them is optimal. In this study, we first characterize a decomposition class specific to CNNs by adopting a flexible graphical notation. The class includes such well-known CNN modules as depthwise separable convolution layers and bottleneck layers, but also previously unknown modules with nonlinear activations. We also experimentally compare the tradeoff between prediction accuracy and time/space complexity for modules found by enumerating all possible decompositions, or by using a neural architecture search. We find some nonlinear decompositions outperform existing ones.","PeriodicalId":72099,"journal":{"name":"Advances in neural information processing systems","volume":"os-53 1","pages":"5553-5563"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advances in neural information processing systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3902/jnns.29.193","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

Tensor decomposition methods are widely used for model compression and fast inference in convolutional neural networks (CNNs). Although many decompositions are conceivable, only CP decomposition and a few others have been applied in practice, and no extensive comparisons have been made between available methods. Previous studies have not determined how many decompositions are available, nor which of them is optimal. In this study, we first characterize a decomposition class specific to CNNs by adopting a flexible graphical notation. The class includes such well-known CNN modules as depthwise separable convolution layers and bottleneck layers, but also previously unknown modules with nonlinear activations. We also experimentally compare the tradeoff between prediction accuracy and time/space complexity for modules found by enumerating all possible decompositions, or by using a neural architecture search. We find some nonlinear decompositions outperform existing ones.
探索卷积神经网络的未探索张量网络分解
张量分解方法在卷积神经网络(cnn)中被广泛用于模型压缩和快速推理。虽然许多分解是可以想象的,但只有CP分解和其他一些分解在实践中得到了应用,并且没有对现有方法进行广泛的比较。以前的研究并没有确定有多少种分解方法可用,也没有确定哪种分解方法是最佳的。在这项研究中,我们首先通过采用灵活的图形符号来表征cnn特有的分解类。该类包括深度可分离卷积层和瓶颈层等众所周知的CNN模块,但也包括以前未知的非线性激活模块。我们还通过实验比较了通过枚举所有可能的分解或使用神经结构搜索找到的模块的预测精度和时间/空间复杂性之间的权衡。我们发现一些非线性分解优于现有的分解。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信