{"title":"卷积神经网络优化与部署的工业级解决方案","authors":"Xinchao Wang, Yongxin Wang, Juan Li","doi":"10.1109/AINIT59027.2023.10212632","DOIUrl":null,"url":null,"abstract":"The deployment of deep learning models into practical production applications has become increasingly important with the rapid development of deep learning in theoretical research. Currently, the deployment of deep learning models faces numerous challenges due to the increasingly large scale and computational requirements of these models, along with the limited storage and computing resources of mobile devices. This paper proposes a high-performance and versatile solution to address the challenges of practical model deployment. This study significantly enhances model inference speed by employing techniques such as DepGraph model pruning, operator fusion, and the NCNN inference framework while reducing the model size and storage overhead.","PeriodicalId":276778,"journal":{"name":"2023 4th International Seminar on Artificial Intelligence, Networking and Information Technology (AINIT)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Industrial-grade Solution for Convolutional Neural Network Optimization and Deployment\",\"authors\":\"Xinchao Wang, Yongxin Wang, Juan Li\",\"doi\":\"10.1109/AINIT59027.2023.10212632\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The deployment of deep learning models into practical production applications has become increasingly important with the rapid development of deep learning in theoretical research. Currently, the deployment of deep learning models faces numerous challenges due to the increasingly large scale and computational requirements of these models, along with the limited storage and computing resources of mobile devices. This paper proposes a high-performance and versatile solution to address the challenges of practical model deployment. This study significantly enhances model inference speed by employing techniques such as DepGraph model pruning, operator fusion, and the NCNN inference framework while reducing the model size and storage overhead.\",\"PeriodicalId\":276778,\"journal\":{\"name\":\"2023 4th International Seminar on Artificial Intelligence, Networking and Information Technology (AINIT)\",\"volume\":\"12 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-06-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 4th International Seminar on Artificial Intelligence, Networking and Information Technology (AINIT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/AINIT59027.2023.10212632\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 4th International Seminar on Artificial Intelligence, Networking and Information Technology (AINIT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AINIT59027.2023.10212632","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An Industrial-grade Solution for Convolutional Neural Network Optimization and Deployment
The deployment of deep learning models into practical production applications has become increasingly important with the rapid development of deep learning in theoretical research. Currently, the deployment of deep learning models faces numerous challenges due to the increasingly large scale and computational requirements of these models, along with the limited storage and computing resources of mobile devices. This paper proposes a high-performance and versatile solution to address the challenges of practical model deployment. This study significantly enhances model inference speed by employing techniques such as DepGraph model pruning, operator fusion, and the NCNN inference framework while reducing the model size and storage overhead.