Lightweight Network for Vietnamese Landmark Recognition based on Knowledge Distillation

V. T. Tran, Nam Le, P. T. Nguyen, Thinh N. Doan
{"title":"Lightweight Network for Vietnamese Landmark Recognition based on Knowledge Distillation","authors":"V. T. Tran, Nam Le, P. T. Nguyen, Thinh N. Doan","doi":"10.1109/NICS51282.2020.9335881","DOIUrl":null,"url":null,"abstract":"In our modern world, smart devices, e.g., mobile phones, IoT devices, have become the norm, leading to a vast increase in demand for a smart-ecosystem. Among other technologies that are being researched and applied, there is a trend of embedding Artificial Intelligence modules on these devices. One of the most challenging problems for embedding on smart devices is maintaining good accuracy while reducing the computational cost and speed. State-of-the-art Deep Convolution Neural Networks cannot run on smart devices due to a lack of resources. The need to find such a model is the motivation for our proposal of a lightweight network for landmark recognition using knowledge distillation. Our purpose is not to create a network with higher accuracy; instead, we try to devise a fast and light neural network while keeping approximately similar accuracy of SOTA models by utilizing knowledge distillation. Our proposed student model achieves a decent result with 7.33% accuracy lower than the teacher SOTA model (91.8%), while decreases the processing time by 73.04%. Our experimental results show promising potential for further explorations and research in knowledge distillation. We have also collected a dataset for Vietnam landmarks for our experiments. This data can be used to train a similar network for Vietnam landmarks recognition or other related purposes.","PeriodicalId":308944,"journal":{"name":"2020 7th NAFOSTED Conference on Information and Computer Science (NICS)","volume":"02 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 7th NAFOSTED Conference on Information and Computer Science (NICS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NICS51282.2020.9335881","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In our modern world, smart devices, e.g., mobile phones, IoT devices, have become the norm, leading to a vast increase in demand for a smart-ecosystem. Among other technologies that are being researched and applied, there is a trend of embedding Artificial Intelligence modules on these devices. One of the most challenging problems for embedding on smart devices is maintaining good accuracy while reducing the computational cost and speed. State-of-the-art Deep Convolution Neural Networks cannot run on smart devices due to a lack of resources. The need to find such a model is the motivation for our proposal of a lightweight network for landmark recognition using knowledge distillation. Our purpose is not to create a network with higher accuracy; instead, we try to devise a fast and light neural network while keeping approximately similar accuracy of SOTA models by utilizing knowledge distillation. Our proposed student model achieves a decent result with 7.33% accuracy lower than the teacher SOTA model (91.8%), while decreases the processing time by 73.04%. Our experimental results show promising potential for further explorations and research in knowledge distillation. We have also collected a dataset for Vietnam landmarks for our experiments. This data can be used to train a similar network for Vietnam landmarks recognition or other related purposes.
基于知识蒸馏的越南地标识别轻量级网络
在我们的现代世界中,智能设备,例如移动电话,物联网设备,已经成为常态,导致对智能生态系统的需求大幅增加。在其他正在研究和应用的技术中,有一种趋势是在这些设备上嵌入人工智能模块。在智能设备上嵌入最具挑战性的问题之一是在降低计算成本和速度的同时保持良好的精度。由于缺乏资源,最先进的深度卷积神经网络无法在智能设备上运行。找到这样一个模型的需要是我们提出使用知识蒸馏进行地标识别的轻量级网络的动机。我们的目的不是创建一个更高精度的网络;相反,我们试图设计一个快速和轻量级的神经网络,同时利用知识蒸馏保持近似于SOTA模型的精度。我们提出的学生模型取得了不错的结果,准确率比教师SOTA模型(91.8%)低7.33%,而处理时间减少了73.04%。我们的实验结果显示了进一步探索和研究知识蒸馏的潜力。我们还为我们的实验收集了越南地标的数据集。这些数据可以用于训练越南地标识别或其他相关目的的类似网络。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信