{"title":"On fault injection approaches for fault tolerance of feedforward neural networks","authors":"Takehiro Ito, I. Takanami","doi":"10.1109/ATS.1997.643927","DOIUrl":null,"url":null,"abstract":"To make a neural network fault-tolerant, Tan et al. proposed a learning algorithm which injects intentionally the snapping of a wire one by one into a network (1992, 1992, 1993). This paper proposes a learning algorithm that injects intentionally stuck-at faults to neurons. Then by computer simulations, we investigate the recognition rate in terms of the number of snapping faults and reliabilities of lines and the learning cycle. The results show that our method is more efficient and useful than the method of Tan et al. Furthermore, we investigate the internal structure in terms of ditribution of correlations between input values of a output neuron for the respective learning methods and show that there is a significant difference of the distributions among the methods.","PeriodicalId":330767,"journal":{"name":"Proceedings Sixth Asian Test Symposium (ATS'97)","volume":"63 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1997-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"24","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings Sixth Asian Test Symposium (ATS'97)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ATS.1997.643927","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 24
Abstract
To make a neural network fault-tolerant, Tan et al. proposed a learning algorithm which injects intentionally the snapping of a wire one by one into a network (1992, 1992, 1993). This paper proposes a learning algorithm that injects intentionally stuck-at faults to neurons. Then by computer simulations, we investigate the recognition rate in terms of the number of snapping faults and reliabilities of lines and the learning cycle. The results show that our method is more efficient and useful than the method of Tan et al. Furthermore, we investigate the internal structure in terms of ditribution of correlations between input values of a output neuron for the respective learning methods and show that there is a significant difference of the distributions among the methods.