使用非理想电阻存储设备的神经网络学习

IF 4.1 Q2 MATERIALS SCIENCE, MULTIDISCIPLINARY
Youngseok Kim, T. Gokmen, H. Miyazoe, P. Solomon, Seyoung Kim, Asit Ray, J. Doevenspeck, R. S. Khan, V. Narayanan, T. Ando
{"title":"使用非理想电阻存储设备的神经网络学习","authors":"Youngseok Kim, T. Gokmen, H. Miyazoe, P. Solomon, Seyoung Kim, Asit Ray, J. Doevenspeck, R. S. Khan, V. Narayanan, T. Ando","doi":"10.3389/fnano.2022.1008266","DOIUrl":null,"url":null,"abstract":"We demonstrate a modified stochastic gradient (Tiki-Taka v2 or TTv2) algorithm for deep learning network training in a cross-bar array architecture based on ReRAM cells. There have been limited discussions on cross-bar arrays for training applications due to the challenges in the switching behavior of nonvolatile memory materials. TTv2 algorithm is known to overcome the device non-idealities for deep learning training. We demonstrate the feasibility of the algorithm for a linear regression task using 1R and 1T1R ReRAM devices. Using the measured device properties, we project the performance of a long short-term memory (LSTM) network with 78 K parameters. We show that TTv2 algorithm relaxes the criteria for symmetric device update response. In addition, further optimization of the algorithm increases noise robustness and significantly reduces the required number of states, thereby drastically improving the model accuracy even with non-ideal devices and achieving the test error close to that of the conventional learning algorithm with an ideal device.","PeriodicalId":34432,"journal":{"name":"Frontiers in Nanotechnology","volume":" ","pages":""},"PeriodicalIF":4.1000,"publicationDate":"2022-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Neural network learning using non-ideal resistive memory devices\",\"authors\":\"Youngseok Kim, T. Gokmen, H. Miyazoe, P. Solomon, Seyoung Kim, Asit Ray, J. Doevenspeck, R. S. Khan, V. Narayanan, T. Ando\",\"doi\":\"10.3389/fnano.2022.1008266\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We demonstrate a modified stochastic gradient (Tiki-Taka v2 or TTv2) algorithm for deep learning network training in a cross-bar array architecture based on ReRAM cells. There have been limited discussions on cross-bar arrays for training applications due to the challenges in the switching behavior of nonvolatile memory materials. TTv2 algorithm is known to overcome the device non-idealities for deep learning training. We demonstrate the feasibility of the algorithm for a linear regression task using 1R and 1T1R ReRAM devices. Using the measured device properties, we project the performance of a long short-term memory (LSTM) network with 78 K parameters. We show that TTv2 algorithm relaxes the criteria for symmetric device update response. In addition, further optimization of the algorithm increases noise robustness and significantly reduces the required number of states, thereby drastically improving the model accuracy even with non-ideal devices and achieving the test error close to that of the conventional learning algorithm with an ideal device.\",\"PeriodicalId\":34432,\"journal\":{\"name\":\"Frontiers in Nanotechnology\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":4.1000,\"publicationDate\":\"2022-10-31\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Frontiers in Nanotechnology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3389/fnano.2022.1008266\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"MATERIALS SCIENCE, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Nanotechnology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/fnano.2022.1008266","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MATERIALS SCIENCE, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

摘要

我们展示了一种改进的随机梯度(Tiki Taka v2或TTv2)算法,用于基于ReRAM单元的横杆阵列结构中的深度学习网络训练。由于非易失性存储器材料的开关行为方面的挑战,关于用于训练应用的横杆阵列的讨论有限。已知TTv2算法可以克服深度学习训练的设备非理想性。我们使用1R和1T1R ReRAM设备证明了该算法用于线性回归任务的可行性。使用测量的设备特性,我们预测了具有78K参数的长短期存储器(LSTM)网络的性能。我们证明了TTv2算法放宽了对称设备更新响应的标准。此外,算法的进一步优化提高了噪声鲁棒性,并显著减少了所需的状态数量,从而即使在非理想设备的情况下也能大幅提高模型精度,并实现了与理想设备的传统学习算法接近的测试误差。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Neural network learning using non-ideal resistive memory devices
We demonstrate a modified stochastic gradient (Tiki-Taka v2 or TTv2) algorithm for deep learning network training in a cross-bar array architecture based on ReRAM cells. There have been limited discussions on cross-bar arrays for training applications due to the challenges in the switching behavior of nonvolatile memory materials. TTv2 algorithm is known to overcome the device non-idealities for deep learning training. We demonstrate the feasibility of the algorithm for a linear regression task using 1R and 1T1R ReRAM devices. Using the measured device properties, we project the performance of a long short-term memory (LSTM) network with 78 K parameters. We show that TTv2 algorithm relaxes the criteria for symmetric device update response. In addition, further optimization of the algorithm increases noise robustness and significantly reduces the required number of states, thereby drastically improving the model accuracy even with non-ideal devices and achieving the test error close to that of the conventional learning algorithm with an ideal device.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Frontiers in Nanotechnology
Frontiers in Nanotechnology Engineering-Electrical and Electronic Engineering
CiteScore
7.10
自引率
0.00%
发文量
96
审稿时长
13 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信