Changyu Zeng, Li Liu, Haocheng Zhao, Yu Zhang, Wei Wang, Ning Cai, Yutao Yue
{"title":"Causal Unstructured Pruning in Linear Networks Using Effective Information","authors":"Changyu Zeng, Li Liu, Haocheng Zhao, Yu Zhang, Wei Wang, Ning Cai, Yutao Yue","doi":"10.1109/CyberC55534.2022.00056","DOIUrl":null,"url":null,"abstract":"Excessive number of parameters in today’s (deep) neural networks demands tremendous computational resources and slows down training speed. The problem also makes it difficult to deploy these neural network models on capability constrained devices such as mobile devices. To address this challenge, we propose an unstructured pruning method that measures the causal structure of neural networks based on effective information (EI). It introduces an intervention to the input and computes the mutual information between the interference and its corresponding output, within a single linear layer measuring the importance of each weight. In the experiments, we found that the sparsity of EI pruning can reach more than 90%. Only 10% of non-zero parameters in the linear layers were needed compared to the benchmark methods without pruning, while ensuring similar level of accuracy and stable training performance in iterative pruning. In addition, as the invariance of the causal structure of the network is exploited, the network after pruning using EI is highly generalizable and interpretable than other methods.","PeriodicalId":234632,"journal":{"name":"2022 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery (CyberC)","volume":"66 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery (CyberC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CyberC55534.2022.00056","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Excessive number of parameters in today’s (deep) neural networks demands tremendous computational resources and slows down training speed. The problem also makes it difficult to deploy these neural network models on capability constrained devices such as mobile devices. To address this challenge, we propose an unstructured pruning method that measures the causal structure of neural networks based on effective information (EI). It introduces an intervention to the input and computes the mutual information between the interference and its corresponding output, within a single linear layer measuring the importance of each weight. In the experiments, we found that the sparsity of EI pruning can reach more than 90%. Only 10% of non-zero parameters in the linear layers were needed compared to the benchmark methods without pruning, while ensuring similar level of accuracy and stable training performance in iterative pruning. In addition, as the invariance of the causal structure of the network is exploited, the network after pruning using EI is highly generalizable and interpretable than other methods.