Khamzul Rifki, Irfan Dwiki Bhaswara, Andi Sulasikin, B. I. Nasution, Y. Nugraha, J. Kanggrawan, M. E. Aminanto
{"title":"雅加达超级应用平台自动收据识别的知识蒸馏","authors":"Khamzul Rifki, Irfan Dwiki Bhaswara, Andi Sulasikin, B. I. Nasution, Y. Nugraha, J. Kanggrawan, M. E. Aminanto","doi":"10.1109/ICITEE56407.2022.9954068","DOIUrl":null,"url":null,"abstract":"Computer vision research has been used in daily applications, such as art, social media app filter, and face recognition. This emergence is because of the usage of the deep learning method in the computer vision domain. Deep learning research has improved many qualities of services for various applications. Starting from recommended until detection systems are now relying on deep learning models. However, currently many models require high computational processing and storage space. Implementing such an extensive network with limited resources on an embedded device or smartphone becomes more challenging. In this study, we focus on developing a model with small computational resources with high accuracy using the knowledge distillation method. We evaluate our model on the public and private datasets of receipt and non-receipt images that we gathered from Badan Pendapatan Daerah, CORD, and Kaggle dataset. After that, we compare it with the regular convolutional neural network (CNN) and pre-trained model. We discovered that knowledge distillation only uses 12% and 5% of the total weight of the CNN and the pre-trained model, respectively. As a result, we see a possibility that knowledge distillation illustrates potential outcomes as a method that could implement for automatic receipt identification in the Jakarta Super App.","PeriodicalId":246279,"journal":{"name":"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Knowledge Distillation for Automatic Receipt Identification in Jakarta Super App Platform\",\"authors\":\"Khamzul Rifki, Irfan Dwiki Bhaswara, Andi Sulasikin, B. I. Nasution, Y. Nugraha, J. Kanggrawan, M. E. Aminanto\",\"doi\":\"10.1109/ICITEE56407.2022.9954068\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Computer vision research has been used in daily applications, such as art, social media app filter, and face recognition. This emergence is because of the usage of the deep learning method in the computer vision domain. Deep learning research has improved many qualities of services for various applications. Starting from recommended until detection systems are now relying on deep learning models. However, currently many models require high computational processing and storage space. Implementing such an extensive network with limited resources on an embedded device or smartphone becomes more challenging. In this study, we focus on developing a model with small computational resources with high accuracy using the knowledge distillation method. We evaluate our model on the public and private datasets of receipt and non-receipt images that we gathered from Badan Pendapatan Daerah, CORD, and Kaggle dataset. After that, we compare it with the regular convolutional neural network (CNN) and pre-trained model. We discovered that knowledge distillation only uses 12% and 5% of the total weight of the CNN and the pre-trained model, respectively. As a result, we see a possibility that knowledge distillation illustrates potential outcomes as a method that could implement for automatic receipt identification in the Jakarta Super App.\",\"PeriodicalId\":246279,\"journal\":{\"name\":\"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)\",\"volume\":\"42 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-10-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICITEE56407.2022.9954068\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICITEE56407.2022.9954068","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Knowledge Distillation for Automatic Receipt Identification in Jakarta Super App Platform
Computer vision research has been used in daily applications, such as art, social media app filter, and face recognition. This emergence is because of the usage of the deep learning method in the computer vision domain. Deep learning research has improved many qualities of services for various applications. Starting from recommended until detection systems are now relying on deep learning models. However, currently many models require high computational processing and storage space. Implementing such an extensive network with limited resources on an embedded device or smartphone becomes more challenging. In this study, we focus on developing a model with small computational resources with high accuracy using the knowledge distillation method. We evaluate our model on the public and private datasets of receipt and non-receipt images that we gathered from Badan Pendapatan Daerah, CORD, and Kaggle dataset. After that, we compare it with the regular convolutional neural network (CNN) and pre-trained model. We discovered that knowledge distillation only uses 12% and 5% of the total weight of the CNN and the pre-trained model, respectively. As a result, we see a possibility that knowledge distillation illustrates potential outcomes as a method that could implement for automatic receipt identification in the Jakarta Super App.