Corey Lammie, Hadjer Benmeziane, William Simon, Elena Ferro, Athanasios Vasilopoulos, Julian Büchel, Manuel Le Gallo, Irem Boybat, Abu Sebastian
{"title":"用于模拟内存中基于计算的加速器的深度学习软件堆栈","authors":"Corey Lammie, Hadjer Benmeziane, William Simon, Elena Ferro, Athanasios Vasilopoulos, Julian Büchel, Manuel Le Gallo, Irem Boybat, Abu Sebastian","doi":"10.1038/s44287-025-00187-1","DOIUrl":null,"url":null,"abstract":"Analogue in-memory computing (AIMC) is an emerging computational paradigm that can efficiently accelerate the key operations in deep learning (DL) inference workloads. Heterogeneous architectures, which integrate both AIMC tiles and digital processing units, have been proposed to enable the end-to-end execution of various deep neural network models. However, developing a software stack for these architectures is challenging, owing to their distinct characteristics — such as the need for extensive or complete weight stationarity and pipelined execution across layers, if maximum performance is to be achieved. Moreover, AIMC tiles are inherently stochastic and hence introduce a combination of stochastic and deterministic noise, which adversely affects accuracy. As a result, existing tools for software stack development are not directly applicable. In this Perspective, we give an overview of the key attributes of DL software stacks and AIMC-based accelerators, outline the challenges associated with designing DL software stacks for AIMC-based accelerators and present opportunities for future research. Analogue in-memory computing (AIMC), with digital processing, forms a useful architecture for performant end-to-end execution of deep neural network models, but requires the development of sophisticated software stacks. This Perspective outlines the challenges in designing deep learning software stacks for AIMC-based accelerators, and suggests directions for future research.","PeriodicalId":501701,"journal":{"name":"Nature Reviews Electrical Engineering","volume":"2 9","pages":"621-633"},"PeriodicalIF":0.0000,"publicationDate":"2025-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Deep learning software stacks for analogue in-memory computing-based accelerators\",\"authors\":\"Corey Lammie, Hadjer Benmeziane, William Simon, Elena Ferro, Athanasios Vasilopoulos, Julian Büchel, Manuel Le Gallo, Irem Boybat, Abu Sebastian\",\"doi\":\"10.1038/s44287-025-00187-1\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Analogue in-memory computing (AIMC) is an emerging computational paradigm that can efficiently accelerate the key operations in deep learning (DL) inference workloads. Heterogeneous architectures, which integrate both AIMC tiles and digital processing units, have been proposed to enable the end-to-end execution of various deep neural network models. However, developing a software stack for these architectures is challenging, owing to their distinct characteristics — such as the need for extensive or complete weight stationarity and pipelined execution across layers, if maximum performance is to be achieved. Moreover, AIMC tiles are inherently stochastic and hence introduce a combination of stochastic and deterministic noise, which adversely affects accuracy. As a result, existing tools for software stack development are not directly applicable. In this Perspective, we give an overview of the key attributes of DL software stacks and AIMC-based accelerators, outline the challenges associated with designing DL software stacks for AIMC-based accelerators and present opportunities for future research. Analogue in-memory computing (AIMC), with digital processing, forms a useful architecture for performant end-to-end execution of deep neural network models, but requires the development of sophisticated software stacks. This Perspective outlines the challenges in designing deep learning software stacks for AIMC-based accelerators, and suggests directions for future research.\",\"PeriodicalId\":501701,\"journal\":{\"name\":\"Nature Reviews Electrical Engineering\",\"volume\":\"2 9\",\"pages\":\"621-633\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-07-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Nature Reviews Electrical Engineering\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.nature.com/articles/s44287-025-00187-1\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Nature Reviews Electrical Engineering","FirstCategoryId":"1085","ListUrlMain":"https://www.nature.com/articles/s44287-025-00187-1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Deep learning software stacks for analogue in-memory computing-based accelerators
Analogue in-memory computing (AIMC) is an emerging computational paradigm that can efficiently accelerate the key operations in deep learning (DL) inference workloads. Heterogeneous architectures, which integrate both AIMC tiles and digital processing units, have been proposed to enable the end-to-end execution of various deep neural network models. However, developing a software stack for these architectures is challenging, owing to their distinct characteristics — such as the need for extensive or complete weight stationarity and pipelined execution across layers, if maximum performance is to be achieved. Moreover, AIMC tiles are inherently stochastic and hence introduce a combination of stochastic and deterministic noise, which adversely affects accuracy. As a result, existing tools for software stack development are not directly applicable. In this Perspective, we give an overview of the key attributes of DL software stacks and AIMC-based accelerators, outline the challenges associated with designing DL software stacks for AIMC-based accelerators and present opportunities for future research. Analogue in-memory computing (AIMC), with digital processing, forms a useful architecture for performant end-to-end execution of deep neural network models, but requires the development of sophisticated software stacks. This Perspective outlines the challenges in designing deep learning software stacks for AIMC-based accelerators, and suggests directions for future research.