Davi Rodrigues, Eleonora Raimondo, Riccardo Tomasello, Mario Carpentieri, Giovanni Finocchio
{"title":"用于部署边缘计算神经形态硬件的磁隧道结设计","authors":"Davi Rodrigues, Eleonora Raimondo, Riccardo Tomasello, Mario Carpentieri, Giovanni Finocchio","doi":"arxiv-2409.02528","DOIUrl":null,"url":null,"abstract":"The electrically readable complex dynamics of robust and scalable magnetic\ntunnel junctions (MTJs) offer promising opportunities for advancing\nneuromorphic computing. In this work, we present an MTJ design with a free\nlayer and two polarizers capable of computing the sigmoidal activation function\nand its gradient at the device level. This design enables both feedforward and\nbackpropagation computations within a single device, extending neuromorphic\ncomputing frameworks previously explored in the literature by introducing the\nability to perform backpropagation directly in hardware. Our algorithm\nimplementation reveals two key findings: (i) the small discrepancies between\nthe MTJ-generated curves and the exact software-generated curves have a\nnegligible impact on the performance of the backpropagation algorithm, (ii) the\ndevice implementation is highly robust to inter-device variation and noise, and\n(iii) the proposed method effectively supports transfer learning and knowledge\ndistillation. To demonstrate this, we evaluated the performance of an edge\ncomputing network using weights from a software-trained model implemented with\nour MTJ design. The results show a minimal loss of accuracy of only 0.1% for\nthe Fashion MNIST dataset and 2% for the CIFAR-100 dataset compared to the\noriginal software implementation. These results highlight the potential of our\nMTJ design for compact, hardware-based neural networks in edge computing\napplications, particularly for transfer learning.","PeriodicalId":501083,"journal":{"name":"arXiv - PHYS - Applied Physics","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A design of magnetic tunnel junctions for the deployment of neuromorphic hardware for edge computing\",\"authors\":\"Davi Rodrigues, Eleonora Raimondo, Riccardo Tomasello, Mario Carpentieri, Giovanni Finocchio\",\"doi\":\"arxiv-2409.02528\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The electrically readable complex dynamics of robust and scalable magnetic\\ntunnel junctions (MTJs) offer promising opportunities for advancing\\nneuromorphic computing. In this work, we present an MTJ design with a free\\nlayer and two polarizers capable of computing the sigmoidal activation function\\nand its gradient at the device level. This design enables both feedforward and\\nbackpropagation computations within a single device, extending neuromorphic\\ncomputing frameworks previously explored in the literature by introducing the\\nability to perform backpropagation directly in hardware. Our algorithm\\nimplementation reveals two key findings: (i) the small discrepancies between\\nthe MTJ-generated curves and the exact software-generated curves have a\\nnegligible impact on the performance of the backpropagation algorithm, (ii) the\\ndevice implementation is highly robust to inter-device variation and noise, and\\n(iii) the proposed method effectively supports transfer learning and knowledge\\ndistillation. To demonstrate this, we evaluated the performance of an edge\\ncomputing network using weights from a software-trained model implemented with\\nour MTJ design. The results show a minimal loss of accuracy of only 0.1% for\\nthe Fashion MNIST dataset and 2% for the CIFAR-100 dataset compared to the\\noriginal software implementation. These results highlight the potential of our\\nMTJ design for compact, hardware-based neural networks in edge computing\\napplications, particularly for transfer learning.\",\"PeriodicalId\":501083,\"journal\":{\"name\":\"arXiv - PHYS - Applied Physics\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - PHYS - Applied Physics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.02528\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - PHYS - Applied Physics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.02528","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A design of magnetic tunnel junctions for the deployment of neuromorphic hardware for edge computing
The electrically readable complex dynamics of robust and scalable magnetic
tunnel junctions (MTJs) offer promising opportunities for advancing
neuromorphic computing. In this work, we present an MTJ design with a free
layer and two polarizers capable of computing the sigmoidal activation function
and its gradient at the device level. This design enables both feedforward and
backpropagation computations within a single device, extending neuromorphic
computing frameworks previously explored in the literature by introducing the
ability to perform backpropagation directly in hardware. Our algorithm
implementation reveals two key findings: (i) the small discrepancies between
the MTJ-generated curves and the exact software-generated curves have a
negligible impact on the performance of the backpropagation algorithm, (ii) the
device implementation is highly robust to inter-device variation and noise, and
(iii) the proposed method effectively supports transfer learning and knowledge
distillation. To demonstrate this, we evaluated the performance of an edge
computing network using weights from a software-trained model implemented with
our MTJ design. The results show a minimal loss of accuracy of only 0.1% for
the Fashion MNIST dataset and 2% for the CIFAR-100 dataset compared to the
original software implementation. These results highlight the potential of our
MTJ design for compact, hardware-based neural networks in edge computing
applications, particularly for transfer learning.