{"title":"A unified DNN weight compression framework using reweighted optimization methods","authors":"Mengchen Fan , Tianyun Zhang , Xiaolong Ma , Jiacheng Guo , Zheng Zhan , Shanglin Zhou , Minghai Qin , Caiwen Ding , Baocheng Geng , Makan Fardad , Yanzhi Wang","doi":"10.1016/j.iswa.2025.200556","DOIUrl":null,"url":null,"abstract":"<div><div>To address the large model sizes and intensive computation requirements of deep neural networks (DNNs), weight pruning techniques have been proposed and generally fall into two categories: static regularization-based pruning and dynamic regularization-based pruning. However, the static method often leads to either complex operations or reduced accuracy, while the dynamic method requires extensive time to adjust parameters to maintain accuracy while achieving effective pruning. In this paper, we propose a unified robustness-aware framework for DNN weight pruning that dynamically updates regularization terms bounded by the designated constraint. This framework can generate both non-structured sparsity and different kinds of structured sparsity, and it incorporates adversarial training to enhance the robustness of the sparse model. We further extend our approach into an integrated framework capable of handling multiple DNN compression tasks. Experimental results show that our proposed method increases the compression rate – up to <span><math><mrow><mn>630</mn><mo>×</mo></mrow></math></span> for LeNet-5, <span><math><mrow><mn>45</mn><mo>×</mo></mrow></math></span> for AlexNet, <span><math><mrow><mn>7</mn><mo>.</mo><mn>2</mn><mo>×</mo></mrow></math></span> for MobileNet, <span><math><mrow><mn>3</mn><mo>.</mo><mn>2</mn><mo>×</mo></mrow></math></span> for ResNet-50 – while also reducing training time and simplifying hyperparameter tuning to a <strong>single penalty parameter</strong>. Additionally, our method improves model robustness by <strong>5.07%</strong> for ResNet-18 and <strong>3.34%</strong> for VGG-16 under a 16×pruning rate, outperforming the state-of-the-art ADMM-based hard constraint method.</div></div>","PeriodicalId":100684,"journal":{"name":"Intelligent Systems with Applications","volume":"27 ","pages":"Article 200556"},"PeriodicalIF":0.0000,"publicationDate":"2025-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Intelligent Systems with Applications","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2667305325000821","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
To address the large model sizes and intensive computation requirements of deep neural networks (DNNs), weight pruning techniques have been proposed and generally fall into two categories: static regularization-based pruning and dynamic regularization-based pruning. However, the static method often leads to either complex operations or reduced accuracy, while the dynamic method requires extensive time to adjust parameters to maintain accuracy while achieving effective pruning. In this paper, we propose a unified robustness-aware framework for DNN weight pruning that dynamically updates regularization terms bounded by the designated constraint. This framework can generate both non-structured sparsity and different kinds of structured sparsity, and it incorporates adversarial training to enhance the robustness of the sparse model. We further extend our approach into an integrated framework capable of handling multiple DNN compression tasks. Experimental results show that our proposed method increases the compression rate – up to for LeNet-5, for AlexNet, for MobileNet, for ResNet-50 – while also reducing training time and simplifying hyperparameter tuning to a single penalty parameter. Additionally, our method improves model robustness by 5.07% for ResNet-18 and 3.34% for VGG-16 under a 16×pruning rate, outperforming the state-of-the-art ADMM-based hard constraint method.