{"title":"基于知识蒸馏和批归一化的低延迟深度峰值神经网络训练","authors":"Thi Diem Tran, K. Le, An Luong Truong Nguyen","doi":"10.1109/CINE56307.2022.10037455","DOIUrl":null,"url":null,"abstract":"Spiking Neural Networks (SNNs) can significantly enhance energy efficiency on neuromorphic hardware due their sparse, biological plausibility and binary event (or spike) driven processing. However, from the non-differentiable nature of a spiking neuron, training high-accuracy and low-latency SNNs is challenging. Recent researches continue to look for ways to improve accuracy and latency. To address these issues in SNNs, we propose a technique that concatenates Knowledge Distillation (KD) and Batch Normalization Through Time (BNTT) method in this study. The BNTT boosts low-latency and low-energy training in SNNs by allowing a neuron to handle the spike rate through various timesteps. The KD approach effectively transfers hidden information from the teacher model to the student network, which converts artificial neural network parameters to SNN weights. This concept allows enriching the performance of SNNs better than the prior technique. Experiments are carried out on the Tiny-ImageNet, CIFAR-10, and CIFAR-100 datasets. on various VGG architectures. We reach top-1 accuracy of 55.67% for ImageNet on VGG-11 and 73.11% for the CIFAR-100 dataset on VGG-16. These results demonstrate that our proposal outperforms earlier converted SNNs in accuracy with only 5 timesteps.","PeriodicalId":336238,"journal":{"name":"2022 5th International Conference on Computational Intelligence and Networks (CINE)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Training Low-Latency Deep Spiking Neural Networks with Knowledge Distillation and Batch Normalization Through Time\",\"authors\":\"Thi Diem Tran, K. Le, An Luong Truong Nguyen\",\"doi\":\"10.1109/CINE56307.2022.10037455\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Spiking Neural Networks (SNNs) can significantly enhance energy efficiency on neuromorphic hardware due their sparse, biological plausibility and binary event (or spike) driven processing. However, from the non-differentiable nature of a spiking neuron, training high-accuracy and low-latency SNNs is challenging. Recent researches continue to look for ways to improve accuracy and latency. To address these issues in SNNs, we propose a technique that concatenates Knowledge Distillation (KD) and Batch Normalization Through Time (BNTT) method in this study. The BNTT boosts low-latency and low-energy training in SNNs by allowing a neuron to handle the spike rate through various timesteps. The KD approach effectively transfers hidden information from the teacher model to the student network, which converts artificial neural network parameters to SNN weights. This concept allows enriching the performance of SNNs better than the prior technique. Experiments are carried out on the Tiny-ImageNet, CIFAR-10, and CIFAR-100 datasets. on various VGG architectures. We reach top-1 accuracy of 55.67% for ImageNet on VGG-11 and 73.11% for the CIFAR-100 dataset on VGG-16. These results demonstrate that our proposal outperforms earlier converted SNNs in accuracy with only 5 timesteps.\",\"PeriodicalId\":336238,\"journal\":{\"name\":\"2022 5th International Conference on Computational Intelligence and Networks (CINE)\",\"volume\":\"17 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 5th International Conference on Computational Intelligence and Networks (CINE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CINE56307.2022.10037455\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 5th International Conference on Computational Intelligence and Networks (CINE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CINE56307.2022.10037455","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Training Low-Latency Deep Spiking Neural Networks with Knowledge Distillation and Batch Normalization Through Time
Spiking Neural Networks (SNNs) can significantly enhance energy efficiency on neuromorphic hardware due their sparse, biological plausibility and binary event (or spike) driven processing. However, from the non-differentiable nature of a spiking neuron, training high-accuracy and low-latency SNNs is challenging. Recent researches continue to look for ways to improve accuracy and latency. To address these issues in SNNs, we propose a technique that concatenates Knowledge Distillation (KD) and Batch Normalization Through Time (BNTT) method in this study. The BNTT boosts low-latency and low-energy training in SNNs by allowing a neuron to handle the spike rate through various timesteps. The KD approach effectively transfers hidden information from the teacher model to the student network, which converts artificial neural network parameters to SNN weights. This concept allows enriching the performance of SNNs better than the prior technique. Experiments are carried out on the Tiny-ImageNet, CIFAR-10, and CIFAR-100 datasets. on various VGG architectures. We reach top-1 accuracy of 55.67% for ImageNet on VGG-11 and 73.11% for the CIFAR-100 dataset on VGG-16. These results demonstrate that our proposal outperforms earlier converted SNNs in accuracy with only 5 timesteps.