Md. Ismail Hossain;Mohammed Rakib;M. M. Lutfe Elahi;Nabeel Mohammed;Shafin Rahman
{"title":"基于循环重叠彩票的卷积神经网络快速剪枝","authors":"Md. Ismail Hossain;Mohammed Rakib;M. M. Lutfe Elahi;Nabeel Mohammed;Shafin Rahman","doi":"10.1109/TAI.2025.3534745","DOIUrl":null,"url":null,"abstract":"Pruning refers to the elimination of trivial weights from neural networks. The sub-networks within an overparameterized model produced after pruning are often called lottery tickets. This research aims to generate winning lottery tickets from a set of lottery tickets that can achieve accuracy similar to that of the original unpruned network. We introduce a novel winning ticket called cyclic overlapping lottery ticket (COLT) by data splitting and cyclic retraining of the pruned network from scratch. We apply a cyclic pruning algorithm that keeps only the overlapping weights of different pruned models trained on different data segments. Our results demonstrate that COLT can achieve similar accuracies (obtained by the unpruned model) while maintaining high sparsities. Based on object recognition and detection tasks, we show that the accuracy of COLT is on par with the winning tickets of the lottery ticket hypothesis and, at times, is better. Moreover, COLTs can be generated using fewer iterations than tickets generated by the popular iterative magnitude pruning method. In addition, we also notice that COLTs generated on large datasets can be transferred to small ones without compromising performance, demonstrating its generalizing capability. We conduct all our experiments on Cifar-10, Cifar-100, TinyImageNet, and ImageNet datasets and report superior performance than the state-of-the-art methods. The codes are available at: <uri>https://github.com/ismail31416/COLT</uri>.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 6","pages":"1664-1678"},"PeriodicalIF":0.0000,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10855806","citationCount":"0","resultStr":"{\"title\":\"COLT: Cyclic Overlapping Lottery Tickets for Faster Pruning of Convolutional Neural Networks\",\"authors\":\"Md. Ismail Hossain;Mohammed Rakib;M. M. Lutfe Elahi;Nabeel Mohammed;Shafin Rahman\",\"doi\":\"10.1109/TAI.2025.3534745\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Pruning refers to the elimination of trivial weights from neural networks. The sub-networks within an overparameterized model produced after pruning are often called lottery tickets. This research aims to generate winning lottery tickets from a set of lottery tickets that can achieve accuracy similar to that of the original unpruned network. We introduce a novel winning ticket called cyclic overlapping lottery ticket (COLT) by data splitting and cyclic retraining of the pruned network from scratch. We apply a cyclic pruning algorithm that keeps only the overlapping weights of different pruned models trained on different data segments. Our results demonstrate that COLT can achieve similar accuracies (obtained by the unpruned model) while maintaining high sparsities. Based on object recognition and detection tasks, we show that the accuracy of COLT is on par with the winning tickets of the lottery ticket hypothesis and, at times, is better. Moreover, COLTs can be generated using fewer iterations than tickets generated by the popular iterative magnitude pruning method. In addition, we also notice that COLTs generated on large datasets can be transferred to small ones without compromising performance, demonstrating its generalizing capability. We conduct all our experiments on Cifar-10, Cifar-100, TinyImageNet, and ImageNet datasets and report superior performance than the state-of-the-art methods. The codes are available at: <uri>https://github.com/ismail31416/COLT</uri>.\",\"PeriodicalId\":73305,\"journal\":{\"name\":\"IEEE transactions on artificial intelligence\",\"volume\":\"6 6\",\"pages\":\"1664-1678\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-01-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10855806\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE transactions on artificial intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10855806/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10855806/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
COLT: Cyclic Overlapping Lottery Tickets for Faster Pruning of Convolutional Neural Networks
Pruning refers to the elimination of trivial weights from neural networks. The sub-networks within an overparameterized model produced after pruning are often called lottery tickets. This research aims to generate winning lottery tickets from a set of lottery tickets that can achieve accuracy similar to that of the original unpruned network. We introduce a novel winning ticket called cyclic overlapping lottery ticket (COLT) by data splitting and cyclic retraining of the pruned network from scratch. We apply a cyclic pruning algorithm that keeps only the overlapping weights of different pruned models trained on different data segments. Our results demonstrate that COLT can achieve similar accuracies (obtained by the unpruned model) while maintaining high sparsities. Based on object recognition and detection tasks, we show that the accuracy of COLT is on par with the winning tickets of the lottery ticket hypothesis and, at times, is better. Moreover, COLTs can be generated using fewer iterations than tickets generated by the popular iterative magnitude pruning method. In addition, we also notice that COLTs generated on large datasets can be transferred to small ones without compromising performance, demonstrating its generalizing capability. We conduct all our experiments on Cifar-10, Cifar-100, TinyImageNet, and ImageNet datasets and report superior performance than the state-of-the-art methods. The codes are available at: https://github.com/ismail31416/COLT.