{"title":"基于片上学习的可合成神经网络的体系结构和算法","authors":"A. Tisan, S. Oniga, B. Attila, G. Ciprian","doi":"10.1109/ISSCS.2007.4292702","DOIUrl":null,"url":null,"abstract":"This paper presents a synthesizable programmable logic blocks architectures, describes the associated formula that makes the blocks to be generic for a backpropagation neural network (NN) with on-chip delta rule learning. The architecture proposed herein takes advantage of distinct datapaths for the forward and backward propagation stages to significantly improve the performance of the learning phase. The architecture is easily scalable and able to cope with arbitrary network sizes with the same hardware.","PeriodicalId":225101,"journal":{"name":"2007 International Symposium on Signals, Circuits and Systems","volume":"37 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":"{\"title\":\"Architecture and Algorithms for Syntetizable Neural Networks with On-Chip Learning\",\"authors\":\"A. Tisan, S. Oniga, B. Attila, G. Ciprian\",\"doi\":\"10.1109/ISSCS.2007.4292702\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper presents a synthesizable programmable logic blocks architectures, describes the associated formula that makes the blocks to be generic for a backpropagation neural network (NN) with on-chip delta rule learning. The architecture proposed herein takes advantage of distinct datapaths for the forward and backward propagation stages to significantly improve the performance of the learning phase. The architecture is easily scalable and able to cope with arbitrary network sizes with the same hardware.\",\"PeriodicalId\":225101,\"journal\":{\"name\":\"2007 International Symposium on Signals, Circuits and Systems\",\"volume\":\"37 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2007-07-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"11\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2007 International Symposium on Signals, Circuits and Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISSCS.2007.4292702\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2007 International Symposium on Signals, Circuits and Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISSCS.2007.4292702","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Architecture and Algorithms for Syntetizable Neural Networks with On-Chip Learning
This paper presents a synthesizable programmable logic blocks architectures, describes the associated formula that makes the blocks to be generic for a backpropagation neural network (NN) with on-chip delta rule learning. The architecture proposed herein takes advantage of distinct datapaths for the forward and backward propagation stages to significantly improve the performance of the learning phase. The architecture is easily scalable and able to cope with arbitrary network sizes with the same hardware.