{"title":"突触交叉条阵列记忆技术比较评估--第 2 部分:设计旋钮和 DNN 精度趋势","authors":"Jeffry Victor, Chunguang Wang, Sumeet K. Gupta","doi":"arxiv-2408.05857","DOIUrl":null,"url":null,"abstract":"Crossbar memory arrays have been touted as the workhorse of in-memory\ncomputing (IMC)-based acceleration of Deep Neural Networks (DNNs), but the\nassociated hardware non-idealities limit their efficacy. To address this,\ncross-layer design solutions that reduce the impact of hardware non-idealities\non DNN accuracy are needed. In Part 1 of this paper, we established the\nco-optimization strategies for various memory technologies and their crossbar\narrays, and conducted a comparative technology evaluation in the context of IMC\nrobustness. In this part, we analyze various design knobs such as array size\nand bit-slice (number of bits per device) and their impact on the performance\nof 8T SRAM, ferroelectric transistor (FeFET), Resistive RAM (ReRAM) and\nspin-orbit-torque magnetic RAM (SOT-MRAM) in the context of inference accuracy\nat 7nm technology node. Further, we study the effect of circuit design\nsolutions such as Partial Wordline Activation (PWA) and custom ADC reference\nlevels that reduce the hardware non-idealities and comparatively analyze the\nresponse of each technology to such accuracy enhancing techniques. Our results\non ResNet-20 (with CIFAR-10) show that PWA increases accuracy by up to 32.56%\nwhile custom ADC reference levels yield up to 31.62% accuracy enhancement. We\nobserve that compared to the other technologies, FeFET, by virtue of its small\nlayout height and high distinguishability of its memory states, is best suited\nfor large arrays. For higher bit-slices and a more complex dataset (ResNet-50\nwith Cifar-100) we found that ReRAM matches the performance of FeFET.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"3 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Comparative Evaluation of Memory Technologies for Synaptic Crossbar Arrays- Part 2: Design Knobs and DNN Accuracy Trends\",\"authors\":\"Jeffry Victor, Chunguang Wang, Sumeet K. Gupta\",\"doi\":\"arxiv-2408.05857\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Crossbar memory arrays have been touted as the workhorse of in-memory\\ncomputing (IMC)-based acceleration of Deep Neural Networks (DNNs), but the\\nassociated hardware non-idealities limit their efficacy. To address this,\\ncross-layer design solutions that reduce the impact of hardware non-idealities\\non DNN accuracy are needed. In Part 1 of this paper, we established the\\nco-optimization strategies for various memory technologies and their crossbar\\narrays, and conducted a comparative technology evaluation in the context of IMC\\nrobustness. In this part, we analyze various design knobs such as array size\\nand bit-slice (number of bits per device) and their impact on the performance\\nof 8T SRAM, ferroelectric transistor (FeFET), Resistive RAM (ReRAM) and\\nspin-orbit-torque magnetic RAM (SOT-MRAM) in the context of inference accuracy\\nat 7nm technology node. Further, we study the effect of circuit design\\nsolutions such as Partial Wordline Activation (PWA) and custom ADC reference\\nlevels that reduce the hardware non-idealities and comparatively analyze the\\nresponse of each technology to such accuracy enhancing techniques. Our results\\non ResNet-20 (with CIFAR-10) show that PWA increases accuracy by up to 32.56%\\nwhile custom ADC reference levels yield up to 31.62% accuracy enhancement. We\\nobserve that compared to the other technologies, FeFET, by virtue of its small\\nlayout height and high distinguishability of its memory states, is best suited\\nfor large arrays. For higher bit-slices and a more complex dataset (ResNet-50\\nwith Cifar-100) we found that ReRAM matches the performance of FeFET.\",\"PeriodicalId\":501168,\"journal\":{\"name\":\"arXiv - CS - Emerging Technologies\",\"volume\":\"3 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Emerging Technologies\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.05857\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Emerging Technologies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.05857","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Comparative Evaluation of Memory Technologies for Synaptic Crossbar Arrays- Part 2: Design Knobs and DNN Accuracy Trends
Crossbar memory arrays have been touted as the workhorse of in-memory
computing (IMC)-based acceleration of Deep Neural Networks (DNNs), but the
associated hardware non-idealities limit their efficacy. To address this,
cross-layer design solutions that reduce the impact of hardware non-idealities
on DNN accuracy are needed. In Part 1 of this paper, we established the
co-optimization strategies for various memory technologies and their crossbar
arrays, and conducted a comparative technology evaluation in the context of IMC
robustness. In this part, we analyze various design knobs such as array size
and bit-slice (number of bits per device) and their impact on the performance
of 8T SRAM, ferroelectric transistor (FeFET), Resistive RAM (ReRAM) and
spin-orbit-torque magnetic RAM (SOT-MRAM) in the context of inference accuracy
at 7nm technology node. Further, we study the effect of circuit design
solutions such as Partial Wordline Activation (PWA) and custom ADC reference
levels that reduce the hardware non-idealities and comparatively analyze the
response of each technology to such accuracy enhancing techniques. Our results
on ResNet-20 (with CIFAR-10) show that PWA increases accuracy by up to 32.56%
while custom ADC reference levels yield up to 31.62% accuracy enhancement. We
observe that compared to the other technologies, FeFET, by virtue of its small
layout height and high distinguishability of its memory states, is best suited
for large arrays. For higher bit-slices and a more complex dataset (ResNet-50
with Cifar-100) we found that ReRAM matches the performance of FeFET.