Efficient Weight Pruning for Optical Neural Networks: When Pruned Weights are Non-Zeros

IF 5.3 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Shuo Zhao;Kun Wu;Xin Li;Ying-Chi Chen
{"title":"Efficient Weight Pruning for Optical Neural Networks: When Pruned Weights are Non-Zeros","authors":"Shuo Zhao;Kun Wu;Xin Li;Ying-Chi Chen","doi":"10.1109/TETCI.2025.3547856","DOIUrl":null,"url":null,"abstract":"Optical neural networks (ONNs) have emerged as a promising solution for energy-efficient deep learning. However, their resource-intensive manufacturing process necessitates efficient methods to streamline ONN architectures without sacrificing their performances. Weight pruning presents a potential remedy. Unlike the conventional neural networks, the pruned weights in ONNs are not necessarily zero in general, thereby making most traditional pruning methods inefficient. In this paper, we propose a novel two-stage pruning method tailored for ONNs. In the first stage, a first-order Taylor expansion of the loss function is applied to effectively identify and prune unimportant weights. To determine the shared value for the pruned weights, a novel optimization method is developed. In the second stage, fine-tuning is further applied to adjust the unpruned weights alongside the shared value of pruned weights. Experimental results on multiple public datasets demonstrate the efficacy of our proposed approach. It achieves superior model compression with minimum loss in accuracy over other conventional pruning techniques.","PeriodicalId":13135,"journal":{"name":"IEEE Transactions on Emerging Topics in Computational Intelligence","volume":"9 5","pages":"3572-3581"},"PeriodicalIF":5.3000,"publicationDate":"2025-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Emerging Topics in Computational Intelligence","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10933573/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Optical neural networks (ONNs) have emerged as a promising solution for energy-efficient deep learning. However, their resource-intensive manufacturing process necessitates efficient methods to streamline ONN architectures without sacrificing their performances. Weight pruning presents a potential remedy. Unlike the conventional neural networks, the pruned weights in ONNs are not necessarily zero in general, thereby making most traditional pruning methods inefficient. In this paper, we propose a novel two-stage pruning method tailored for ONNs. In the first stage, a first-order Taylor expansion of the loss function is applied to effectively identify and prune unimportant weights. To determine the shared value for the pruned weights, a novel optimization method is developed. In the second stage, fine-tuning is further applied to adjust the unpruned weights alongside the shared value of pruned weights. Experimental results on multiple public datasets demonstrate the efficacy of our proposed approach. It achieves superior model compression with minimum loss in accuracy over other conventional pruning techniques.
光神经网络的有效权值剪枝:当剪枝权值非零时
光神经网络(ONNs)已经成为一种很有前途的节能深度学习解决方案。然而,它们的资源密集型制造过程需要有效的方法来简化ONN架构,而不牺牲其性能。重量修剪提供了一个潜在的补救措施。与传统神经网络不同,网络的剪枝权值通常不一定为零,这使得大多数传统的剪枝方法效率低下。本文提出了一种针对网络的两阶段剪枝方法。在第一阶段,利用损失函数的一阶泰勒展开式来有效地识别和修剪不重要的权重。为了确定剪枝权值的共享值,提出了一种新的优化方法。在第二阶段,进一步应用微调来调整未修剪的权重以及修剪后的权重的共享值。在多个公共数据集上的实验结果证明了该方法的有效性。与其他传统修剪技术相比,它以最小的精度损失实现了卓越的模型压缩。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
10.30
自引率
7.50%
发文量
147
期刊介绍: The IEEE Transactions on Emerging Topics in Computational Intelligence (TETCI) publishes original articles on emerging aspects of computational intelligence, including theory, applications, and surveys. TETCI is an electronics only publication. TETCI publishes six issues per year. Authors are encouraged to submit manuscripts in any emerging topic in computational intelligence, especially nature-inspired computing topics not covered by other IEEE Computational Intelligence Society journals. A few such illustrative examples are glial cell networks, computational neuroscience, Brain Computer Interface, ambient intelligence, non-fuzzy computing with words, artificial life, cultural learning, artificial endocrine networks, social reasoning, artificial hormone networks, computational intelligence for the IoT and Smart-X technologies.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信