{"title":"基于reram的高效节能神经网络加速器懒引擎研究","authors":"Wei-Yi Yang, Ya-Shu Chen, Jinqi Xiao","doi":"10.1109/INDIN51773.2022.9976171","DOIUrl":null,"url":null,"abstract":"Resistive random-access memory (ReRAM) has been explored to be a promising solution to accelerate the inference of deep neural networks at the embedded systems by performing computations in memory. To reduce the latency of the neural network, all the pre-trained weights are pre-programmed in ReRAM cells as device resistance for the inference phase. However, the system utilization is decreased by the data dependency of the deployed neural networks and results in low energy efficiency. In this work, we propose a Lazy Engine for providing high utilization and energy-efficient ReRAM-based accelerators. Instead of avoiding idle time by applying ReRAM crossbar duplication, Lazy Engine delays the start time of the vector-matrix multiplication operations, with run-time programming overhead consideration, to reclaim idle time for energy efficiency while improving resource utilization. The experimental results show that Lazy Engine achieves up to 77% and 96% improvement in resource utilization and energy saving compared to state-of-the-art ReRAM-based accelerators.","PeriodicalId":359190,"journal":{"name":"2022 IEEE 20th International Conference on Industrial Informatics (INDIN)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Lazy Engine for High-utilization and Energy-efficient ReRAM-based Neural Network Accelerator\",\"authors\":\"Wei-Yi Yang, Ya-Shu Chen, Jinqi Xiao\",\"doi\":\"10.1109/INDIN51773.2022.9976171\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Resistive random-access memory (ReRAM) has been explored to be a promising solution to accelerate the inference of deep neural networks at the embedded systems by performing computations in memory. To reduce the latency of the neural network, all the pre-trained weights are pre-programmed in ReRAM cells as device resistance for the inference phase. However, the system utilization is decreased by the data dependency of the deployed neural networks and results in low energy efficiency. In this work, we propose a Lazy Engine for providing high utilization and energy-efficient ReRAM-based accelerators. Instead of avoiding idle time by applying ReRAM crossbar duplication, Lazy Engine delays the start time of the vector-matrix multiplication operations, with run-time programming overhead consideration, to reclaim idle time for energy efficiency while improving resource utilization. The experimental results show that Lazy Engine achieves up to 77% and 96% improvement in resource utilization and energy saving compared to state-of-the-art ReRAM-based accelerators.\",\"PeriodicalId\":359190,\"journal\":{\"name\":\"2022 IEEE 20th International Conference on Industrial Informatics (INDIN)\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-07-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 20th International Conference on Industrial Informatics (INDIN)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/INDIN51773.2022.9976171\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 20th International Conference on Industrial Informatics (INDIN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/INDIN51773.2022.9976171","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Lazy Engine for High-utilization and Energy-efficient ReRAM-based Neural Network Accelerator
Resistive random-access memory (ReRAM) has been explored to be a promising solution to accelerate the inference of deep neural networks at the embedded systems by performing computations in memory. To reduce the latency of the neural network, all the pre-trained weights are pre-programmed in ReRAM cells as device resistance for the inference phase. However, the system utilization is decreased by the data dependency of the deployed neural networks and results in low energy efficiency. In this work, we propose a Lazy Engine for providing high utilization and energy-efficient ReRAM-based accelerators. Instead of avoiding idle time by applying ReRAM crossbar duplication, Lazy Engine delays the start time of the vector-matrix multiplication operations, with run-time programming overhead consideration, to reclaim idle time for energy efficiency while improving resource utilization. The experimental results show that Lazy Engine achieves up to 77% and 96% improvement in resource utilization and energy saving compared to state-of-the-art ReRAM-based accelerators.