Priscila C. Holanda, C. Reinbrecht, G. Bontorin, Vitor V. Bandeira, R. Reis
{"title":"DHyANA:基于noc的神经网络硬件架构","authors":"Priscila C. Holanda, C. Reinbrecht, G. Bontorin, Vitor V. Bandeira, R. Reis","doi":"10.1109/ICECS.2016.7841161","DOIUrl":null,"url":null,"abstract":"Understanding and modeling the brain is one of the key scientific challenges in the twenty-first century, and a grown effort is rising on a global scale. Due to its high parallelism, the hardware implementation of large-scale spiking neural networks (SNNs) promises superior execution speed compared to sequential software approaches. Such systems can significantly benefit from the use of networks-on-chip(NoC), as they scale very well concerning area, performance, power/energy consumption, and overall design effort. We developed a hierarchical network-on-chip for a hardware SNN architecture to improve the communication and scalability of the system. The architecture was implemented in an Altera Stratix IV FPGA, and a logic synthesis was performed to evaluate the system, achieving an area of 0.23mm2 and a power dissipation of 147mW for a 256 neurons implementation.","PeriodicalId":205556,"journal":{"name":"2016 IEEE International Conference on Electronics, Circuits and Systems (ICECS)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":"{\"title\":\"DHyANA: A NoC-based neural network hardware architecture\",\"authors\":\"Priscila C. Holanda, C. Reinbrecht, G. Bontorin, Vitor V. Bandeira, R. Reis\",\"doi\":\"10.1109/ICECS.2016.7841161\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Understanding and modeling the brain is one of the key scientific challenges in the twenty-first century, and a grown effort is rising on a global scale. Due to its high parallelism, the hardware implementation of large-scale spiking neural networks (SNNs) promises superior execution speed compared to sequential software approaches. Such systems can significantly benefit from the use of networks-on-chip(NoC), as they scale very well concerning area, performance, power/energy consumption, and overall design effort. We developed a hierarchical network-on-chip for a hardware SNN architecture to improve the communication and scalability of the system. The architecture was implemented in an Altera Stratix IV FPGA, and a logic synthesis was performed to evaluate the system, achieving an area of 0.23mm2 and a power dissipation of 147mW for a 256 neurons implementation.\",\"PeriodicalId\":205556,\"journal\":{\"name\":\"2016 IEEE International Conference on Electronics, Circuits and Systems (ICECS)\",\"volume\":\"27 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"10\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 IEEE International Conference on Electronics, Circuits and Systems (ICECS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICECS.2016.7841161\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE International Conference on Electronics, Circuits and Systems (ICECS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICECS.2016.7841161","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 10
摘要
对大脑的理解和建模是21世纪最重要的科学挑战之一,在全球范围内,这方面的努力正在不断加强。由于其高并行性,与顺序软件方法相比,大规模峰值神经网络(snn)的硬件实现保证了更高的执行速度。这样的系统可以从使用片上网络(NoC)中获得显著的好处,因为它们在面积、性能、功耗/能耗和总体设计方面都可以很好地扩展。我们为硬件SNN架构开发了一种分层片上网络,以改善系统的通信和可扩展性。该架构在Altera Stratix IV FPGA上实现,并进行了逻辑综合来评估系统,实现了256个神经元的面积为0.23mm2,功耗为147mW。
DHyANA: A NoC-based neural network hardware architecture
Understanding and modeling the brain is one of the key scientific challenges in the twenty-first century, and a grown effort is rising on a global scale. Due to its high parallelism, the hardware implementation of large-scale spiking neural networks (SNNs) promises superior execution speed compared to sequential software approaches. Such systems can significantly benefit from the use of networks-on-chip(NoC), as they scale very well concerning area, performance, power/energy consumption, and overall design effort. We developed a hierarchical network-on-chip for a hardware SNN architecture to improve the communication and scalability of the system. The architecture was implemented in an Altera Stratix IV FPGA, and a logic synthesis was performed to evaluate the system, achieving an area of 0.23mm2 and a power dissipation of 147mW for a 256 neurons implementation.