{"title":"cimax编译器:面向异构内存计算平台的端到端人工神经网络编译器","authors":"Chen Yang, Yawei Wang, Lei Wu, Xiang Qiu","doi":"10.1109/WCCCT56755.2023.10052488","DOIUrl":null,"url":null,"abstract":"As artificial neural network (ANN) being widely adopted in edge applications, such as voice recognition, and face detection, etc. Computing-In-Memory (CIM) accelerators have received much attention because they are faster and more energy efficient. However, existing compilers are mainly applicable for traditional backend devices rather than emerging hardware architectures like CIM accelerator. In this paper, we propose an end-to-end neural network compiler, named CIMAX-Compiler. It can convert an ANN model in Open Neural Network Exchange (ONNX) format to executable codes, which runs on a heterogeneous embedded system composed of an MCU and a CIM accelerator. CIMAX-Compiler greatly reduces model deployment effort from tens of engineer-hours to less than a second. In addition, we applied several optimization techniques such as operator fusion and convolution compression to further improve compiled code performance. Experimental results show that the optimized code can speed up ANN model inference performance by more than 2× compared with the base-line layer to layer implementation.","PeriodicalId":112978,"journal":{"name":"2023 6th World Conference on Computing and Communication Technologies (WCCCT)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"CIMAX-Compiler: An End-to-End ANN Compiler for Heterogeneous Computing-in-Memory Platform\",\"authors\":\"Chen Yang, Yawei Wang, Lei Wu, Xiang Qiu\",\"doi\":\"10.1109/WCCCT56755.2023.10052488\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"As artificial neural network (ANN) being widely adopted in edge applications, such as voice recognition, and face detection, etc. Computing-In-Memory (CIM) accelerators have received much attention because they are faster and more energy efficient. However, existing compilers are mainly applicable for traditional backend devices rather than emerging hardware architectures like CIM accelerator. In this paper, we propose an end-to-end neural network compiler, named CIMAX-Compiler. It can convert an ANN model in Open Neural Network Exchange (ONNX) format to executable codes, which runs on a heterogeneous embedded system composed of an MCU and a CIM accelerator. CIMAX-Compiler greatly reduces model deployment effort from tens of engineer-hours to less than a second. In addition, we applied several optimization techniques such as operator fusion and convolution compression to further improve compiled code performance. Experimental results show that the optimized code can speed up ANN model inference performance by more than 2× compared with the base-line layer to layer implementation.\",\"PeriodicalId\":112978,\"journal\":{\"name\":\"2023 6th World Conference on Computing and Communication Technologies (WCCCT)\",\"volume\":\"29 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 6th World Conference on Computing and Communication Technologies (WCCCT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/WCCCT56755.2023.10052488\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 6th World Conference on Computing and Communication Technologies (WCCCT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WCCCT56755.2023.10052488","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
CIMAX-Compiler: An End-to-End ANN Compiler for Heterogeneous Computing-in-Memory Platform
As artificial neural network (ANN) being widely adopted in edge applications, such as voice recognition, and face detection, etc. Computing-In-Memory (CIM) accelerators have received much attention because they are faster and more energy efficient. However, existing compilers are mainly applicable for traditional backend devices rather than emerging hardware architectures like CIM accelerator. In this paper, we propose an end-to-end neural network compiler, named CIMAX-Compiler. It can convert an ANN model in Open Neural Network Exchange (ONNX) format to executable codes, which runs on a heterogeneous embedded system composed of an MCU and a CIM accelerator. CIMAX-Compiler greatly reduces model deployment effort from tens of engineer-hours to less than a second. In addition, we applied several optimization techniques such as operator fusion and convolution compression to further improve compiled code performance. Experimental results show that the optimized code can speed up ANN model inference performance by more than 2× compared with the base-line layer to layer implementation.