基于注意力的低复杂度卷积神经网络层剪枝新方法

IF 6.8 Q1 AUTOMATION & CONTROL SYSTEMS
Md. Bipul Hossain, Na Gong, Mohamed Shaban
{"title":"基于注意力的低复杂度卷积神经网络层剪枝新方法","authors":"Md. Bipul Hossain,&nbsp;Na Gong,&nbsp;Mohamed Shaban","doi":"10.1002/aisy.202400161","DOIUrl":null,"url":null,"abstract":"<p>Deep learning (DL) has been very successful for classifying images, detecting targets, and segmenting regions in high-resolution images such as whole slide histopathology images. However, analysis of such high-resolution images requires very high DL complexity. Several AI optimization techniques have been recently proposed that aim at reducing the complexity of deep neural networks and hence expedite their execution and eventually allow the use of low-power, low-cost computing devices with limited computation and memory resources. These methods include parameter pruning and sharing, quantization, knowledge distillation, low-rank approximation, and resource efficient architectures. Rather than pruning network structures including filters, layers, and blocks of layers based on a manual selection of a significance metric such as <i>l</i>1<i>-</i>norm and <i>l</i>2<i>-</i>norm of the filter kernels, novel highly efficient AI-driven DL optimization algorithms using variations of the squeeze and excitation in order to prune filters and layers of deep models such as VGG-16 as well as eliminate filters and blocks of residual networks such as ResNet-56 are introduced. The proposed techniques achieve significantly higher reduction in the number of learning parameters, the number of floating point operations, and memory space as compared to the-state-of-the-art methods.</p>","PeriodicalId":93858,"journal":{"name":"Advanced intelligent systems (Weinheim an der Bergstrasse, Germany)","volume":"6 11","pages":""},"PeriodicalIF":6.8000,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aisy.202400161","citationCount":"0","resultStr":"{\"title\":\"A Novel Attention-Based Layer Pruning Approach for Low-Complexity Convolutional Neural Networks\",\"authors\":\"Md. Bipul Hossain,&nbsp;Na Gong,&nbsp;Mohamed Shaban\",\"doi\":\"10.1002/aisy.202400161\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Deep learning (DL) has been very successful for classifying images, detecting targets, and segmenting regions in high-resolution images such as whole slide histopathology images. However, analysis of such high-resolution images requires very high DL complexity. Several AI optimization techniques have been recently proposed that aim at reducing the complexity of deep neural networks and hence expedite their execution and eventually allow the use of low-power, low-cost computing devices with limited computation and memory resources. These methods include parameter pruning and sharing, quantization, knowledge distillation, low-rank approximation, and resource efficient architectures. Rather than pruning network structures including filters, layers, and blocks of layers based on a manual selection of a significance metric such as <i>l</i>1<i>-</i>norm and <i>l</i>2<i>-</i>norm of the filter kernels, novel highly efficient AI-driven DL optimization algorithms using variations of the squeeze and excitation in order to prune filters and layers of deep models such as VGG-16 as well as eliminate filters and blocks of residual networks such as ResNet-56 are introduced. The proposed techniques achieve significantly higher reduction in the number of learning parameters, the number of floating point operations, and memory space as compared to the-state-of-the-art methods.</p>\",\"PeriodicalId\":93858,\"journal\":{\"name\":\"Advanced intelligent systems (Weinheim an der Bergstrasse, Germany)\",\"volume\":\"6 11\",\"pages\":\"\"},\"PeriodicalIF\":6.8000,\"publicationDate\":\"2024-06-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aisy.202400161\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Advanced intelligent systems (Weinheim an der Bergstrasse, Germany)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/aisy.202400161\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advanced intelligent systems (Weinheim an der Bergstrasse, Germany)","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/aisy.202400161","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

摘要

深度学习(DL)在高分辨率图像(如整张组织病理学图像)的图像分类、目标检测和区域分割方面非常成功。然而,分析此类高分辨率图像需要非常高的深度学习复杂度。最近提出了几种人工智能优化技术,旨在降低深度神经网络的复杂性,从而加快其执行速度,并最终允许使用计算和内存资源有限的低功耗、低成本计算设备。这些方法包括参数剪枝和共享、量化、知识提炼、低秩逼近和资源高效架构。与基于人工选择重要度量(如滤波器内核的 l1-norm 和 l2-norm)来修剪包括滤波器、层和层块在内的网络结构不同,本文引入了新型高效人工智能驱动的 DL 优化算法,利用挤压和激励的变化来修剪深度模型(如 VGG-16)的滤波器和层,以及消除残差网络(如 ResNet-56)的滤波器和块。与最先进的方法相比,所提出的技术大大减少了学习参数数量、浮点运算次数和内存空间。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

A Novel Attention-Based Layer Pruning Approach for Low-Complexity Convolutional Neural Networks

A Novel Attention-Based Layer Pruning Approach for Low-Complexity Convolutional Neural Networks

Deep learning (DL) has been very successful for classifying images, detecting targets, and segmenting regions in high-resolution images such as whole slide histopathology images. However, analysis of such high-resolution images requires very high DL complexity. Several AI optimization techniques have been recently proposed that aim at reducing the complexity of deep neural networks and hence expedite their execution and eventually allow the use of low-power, low-cost computing devices with limited computation and memory resources. These methods include parameter pruning and sharing, quantization, knowledge distillation, low-rank approximation, and resource efficient architectures. Rather than pruning network structures including filters, layers, and blocks of layers based on a manual selection of a significance metric such as l1-norm and l2-norm of the filter kernels, novel highly efficient AI-driven DL optimization algorithms using variations of the squeeze and excitation in order to prune filters and layers of deep models such as VGG-16 as well as eliminate filters and blocks of residual networks such as ResNet-56 are introduced. The proposed techniques achieve significantly higher reduction in the number of learning parameters, the number of floating point operations, and memory space as compared to the-state-of-the-art methods.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
1.30
自引率
0.00%
发文量
0
审稿时长
4 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信