雅加达超级应用平台自动收据识别的知识蒸馏

Khamzul Rifki, Irfan Dwiki Bhaswara, Andi Sulasikin, B. I. Nasution, Y. Nugraha, J. Kanggrawan, M. E. Aminanto
{"title":"雅加达超级应用平台自动收据识别的知识蒸馏","authors":"Khamzul Rifki, Irfan Dwiki Bhaswara, Andi Sulasikin, B. I. Nasution, Y. Nugraha, J. Kanggrawan, M. E. Aminanto","doi":"10.1109/ICITEE56407.2022.9954068","DOIUrl":null,"url":null,"abstract":"Computer vision research has been used in daily applications, such as art, social media app filter, and face recognition. This emergence is because of the usage of the deep learning method in the computer vision domain. Deep learning research has improved many qualities of services for various applications. Starting from recommended until detection systems are now relying on deep learning models. However, currently many models require high computational processing and storage space. Implementing such an extensive network with limited resources on an embedded device or smartphone becomes more challenging. In this study, we focus on developing a model with small computational resources with high accuracy using the knowledge distillation method. We evaluate our model on the public and private datasets of receipt and non-receipt images that we gathered from Badan Pendapatan Daerah, CORD, and Kaggle dataset. After that, we compare it with the regular convolutional neural network (CNN) and pre-trained model. We discovered that knowledge distillation only uses 12% and 5% of the total weight of the CNN and the pre-trained model, respectively. As a result, we see a possibility that knowledge distillation illustrates potential outcomes as a method that could implement for automatic receipt identification in the Jakarta Super App.","PeriodicalId":246279,"journal":{"name":"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Knowledge Distillation for Automatic Receipt Identification in Jakarta Super App Platform\",\"authors\":\"Khamzul Rifki, Irfan Dwiki Bhaswara, Andi Sulasikin, B. I. Nasution, Y. Nugraha, J. Kanggrawan, M. E. Aminanto\",\"doi\":\"10.1109/ICITEE56407.2022.9954068\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Computer vision research has been used in daily applications, such as art, social media app filter, and face recognition. This emergence is because of the usage of the deep learning method in the computer vision domain. Deep learning research has improved many qualities of services for various applications. Starting from recommended until detection systems are now relying on deep learning models. However, currently many models require high computational processing and storage space. Implementing such an extensive network with limited resources on an embedded device or smartphone becomes more challenging. In this study, we focus on developing a model with small computational resources with high accuracy using the knowledge distillation method. We evaluate our model on the public and private datasets of receipt and non-receipt images that we gathered from Badan Pendapatan Daerah, CORD, and Kaggle dataset. After that, we compare it with the regular convolutional neural network (CNN) and pre-trained model. We discovered that knowledge distillation only uses 12% and 5% of the total weight of the CNN and the pre-trained model, respectively. As a result, we see a possibility that knowledge distillation illustrates potential outcomes as a method that could implement for automatic receipt identification in the Jakarta Super App.\",\"PeriodicalId\":246279,\"journal\":{\"name\":\"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)\",\"volume\":\"42 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-10-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICITEE56407.2022.9954068\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 14th International Conference on Information Technology and Electrical Engineering (ICITEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICITEE56407.2022.9954068","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

计算机视觉研究已经应用于日常应用,如艺术、社交媒体应用程序过滤器和人脸识别。这种出现是因为深度学习方法在计算机视觉领域的应用。深度学习研究提高了各种应用的服务质量。从推荐到检测系统现在依赖于深度学习模型。然而,目前许多模型对计算处理和存储空间的要求很高。在嵌入式设备或智能手机上使用有限的资源实现如此广泛的网络变得更具挑战性。在本研究中,我们着重于利用知识蒸馏方法开发一个计算资源少、精度高的模型。我们在从Badan Pendapatan Daerah、CORD和Kaggle数据集收集的收据和非收据图像的公共和私人数据集上评估我们的模型。之后,我们将其与常规卷积神经网络(CNN)和预训练模型进行比较。我们发现知识蒸馏分别只使用了CNN和预训练模型总权重的12%和5%。因此,我们看到了一种可能性,即知识蒸馏说明了作为一种可以在雅加达超级应用程序中实现自动收据识别的方法的潜在结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Knowledge Distillation for Automatic Receipt Identification in Jakarta Super App Platform
Computer vision research has been used in daily applications, such as art, social media app filter, and face recognition. This emergence is because of the usage of the deep learning method in the computer vision domain. Deep learning research has improved many qualities of services for various applications. Starting from recommended until detection systems are now relying on deep learning models. However, currently many models require high computational processing and storage space. Implementing such an extensive network with limited resources on an embedded device or smartphone becomes more challenging. In this study, we focus on developing a model with small computational resources with high accuracy using the knowledge distillation method. We evaluate our model on the public and private datasets of receipt and non-receipt images that we gathered from Badan Pendapatan Daerah, CORD, and Kaggle dataset. After that, we compare it with the regular convolutional neural network (CNN) and pre-trained model. We discovered that knowledge distillation only uses 12% and 5% of the total weight of the CNN and the pre-trained model, respectively. As a result, we see a possibility that knowledge distillation illustrates potential outcomes as a method that could implement for automatic receipt identification in the Jakarta Super App.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信