{"title":"一种有效支持人工神经网络的智能存储器架构","authors":"K. Großpietsch, J. Büddefeld","doi":"10.1109/EMPDP.2001.905074","DOIUrl":null,"url":null,"abstract":"A \"smart memory\" approach is presented, i.e. the new architecture is achieved by extending the functionality of a conventional RAM structure. The architecture additionally contains two innovative features: To every word cell of w bits, a small q bits wide ALU is associated; and by means of extending the memory decoder, multiple access to certain sets of word cells within the memory as well as activation of their ALUs is possible. It is shown that based on these features, the standard numerical problem of adding up the m components of a vector of dimension m, in the new architecture can be carried out in a time complexity of O(square root(m)). For the execution of artificial neural nets, especially the on-line recognition of patterns mainly depends on the time-efficient efficient execution of weighted sums. It is shown that in our architecture, these weighted sums can be computed quite efficiently. The computation time is highly superior to the time complexity on sequential von Neumann machines. In addition, we show that if requested, the training mode of a neural net can also be significantly be speeded up. This is achieved by means of a simple crossbar switch which can be modularly added to the array of memory chips.","PeriodicalId":262971,"journal":{"name":"Proceedings Ninth Euromicro Workshop on Parallel and Distributed Processing","volume":"269 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2001-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"A smart memory architecture for the efficient support of artificial neural nets\",\"authors\":\"K. Großpietsch, J. Büddefeld\",\"doi\":\"10.1109/EMPDP.2001.905074\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A \\\"smart memory\\\" approach is presented, i.e. the new architecture is achieved by extending the functionality of a conventional RAM structure. The architecture additionally contains two innovative features: To every word cell of w bits, a small q bits wide ALU is associated; and by means of extending the memory decoder, multiple access to certain sets of word cells within the memory as well as activation of their ALUs is possible. It is shown that based on these features, the standard numerical problem of adding up the m components of a vector of dimension m, in the new architecture can be carried out in a time complexity of O(square root(m)). For the execution of artificial neural nets, especially the on-line recognition of patterns mainly depends on the time-efficient efficient execution of weighted sums. It is shown that in our architecture, these weighted sums can be computed quite efficiently. The computation time is highly superior to the time complexity on sequential von Neumann machines. In addition, we show that if requested, the training mode of a neural net can also be significantly be speeded up. This is achieved by means of a simple crossbar switch which can be modularly added to the array of memory chips.\",\"PeriodicalId\":262971,\"journal\":{\"name\":\"Proceedings Ninth Euromicro Workshop on Parallel and Distributed Processing\",\"volume\":\"269 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2001-02-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings Ninth Euromicro Workshop on Parallel and Distributed Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/EMPDP.2001.905074\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings Ninth Euromicro Workshop on Parallel and Distributed Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/EMPDP.2001.905074","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A smart memory architecture for the efficient support of artificial neural nets
A "smart memory" approach is presented, i.e. the new architecture is achieved by extending the functionality of a conventional RAM structure. The architecture additionally contains two innovative features: To every word cell of w bits, a small q bits wide ALU is associated; and by means of extending the memory decoder, multiple access to certain sets of word cells within the memory as well as activation of their ALUs is possible. It is shown that based on these features, the standard numerical problem of adding up the m components of a vector of dimension m, in the new architecture can be carried out in a time complexity of O(square root(m)). For the execution of artificial neural nets, especially the on-line recognition of patterns mainly depends on the time-efficient efficient execution of weighted sums. It is shown that in our architecture, these weighted sums can be computed quite efficiently. The computation time is highly superior to the time complexity on sequential von Neumann machines. In addition, we show that if requested, the training mode of a neural net can also be significantly be speeded up. This is achieved by means of a simple crossbar switch which can be modularly added to the array of memory chips.