{"title":"基于嵌入式设备的多人姿态估计","authors":"Zhipeng Ma, Dawei Tian, Ming Zhang, Dingxin He","doi":"10.1109/ICHCI51889.2020.00020","DOIUrl":null,"url":null,"abstract":"Tasks based on deep learning have make great progress in accuracy with the development of deep learning in recent years. However, algorithms base on deep learning are both computationally intensive and memory intensive, makes it difficult to deploy on resource-constrained devices. Therefore, it is worth studying how to compress the architecture of existing network while achieving a comparable performance, so that the model can be deployed on devices with limited resources. This paper performs model compression and acceleration in multi-person pose estimation by replacing the feature extraction network, parameter pruning and knowledge distillation, achieve a $2.6\\times$ MACs reduction and $2\\times$ acceleration but only 25% drop in accuracy compared with a lightweight model. The compressed model can be deployed on the embedded device Jetson Nano with a 12 FPS inference speed.","PeriodicalId":355427,"journal":{"name":"2020 International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Multi-Person Pose Estimation on Embedded Device\",\"authors\":\"Zhipeng Ma, Dawei Tian, Ming Zhang, Dingxin He\",\"doi\":\"10.1109/ICHCI51889.2020.00020\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Tasks based on deep learning have make great progress in accuracy with the development of deep learning in recent years. However, algorithms base on deep learning are both computationally intensive and memory intensive, makes it difficult to deploy on resource-constrained devices. Therefore, it is worth studying how to compress the architecture of existing network while achieving a comparable performance, so that the model can be deployed on devices with limited resources. This paper performs model compression and acceleration in multi-person pose estimation by replacing the feature extraction network, parameter pruning and knowledge distillation, achieve a $2.6\\\\times$ MACs reduction and $2\\\\times$ acceleration but only 25% drop in accuracy compared with a lightweight model. The compressed model can be deployed on the embedded device Jetson Nano with a 12 FPS inference speed.\",\"PeriodicalId\":355427,\"journal\":{\"name\":\"2020 International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI)\",\"volume\":\"44 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICHCI51889.2020.00020\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICHCI51889.2020.00020","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Tasks based on deep learning have make great progress in accuracy with the development of deep learning in recent years. However, algorithms base on deep learning are both computationally intensive and memory intensive, makes it difficult to deploy on resource-constrained devices. Therefore, it is worth studying how to compress the architecture of existing network while achieving a comparable performance, so that the model can be deployed on devices with limited resources. This paper performs model compression and acceleration in multi-person pose estimation by replacing the feature extraction network, parameter pruning and knowledge distillation, achieve a $2.6\times$ MACs reduction and $2\times$ acceleration but only 25% drop in accuracy compared with a lightweight model. The compressed model can be deployed on the embedded device Jetson Nano with a 12 FPS inference speed.