图像分类自主学习平台的研究与实现

Shiyu Zhao, Menghua Jiang, Zengwen Li, Changxue Chen, Liang Song
{"title":"图像分类自主学习平台的研究与实现","authors":"Shiyu Zhao, Menghua Jiang, Zengwen Li, Changxue Chen, Liang Song","doi":"10.1109/INSAI56792.2022.00039","DOIUrl":null,"url":null,"abstract":"The image classification task aims to automatically classify image content based on machine learning methods. This task is a basic task in the field of computer vision, which has broad application prospects and great research value. At present, in the case of large-scale corpus annotation, mainstream image classification algorithms based on deep learning have been able to obtain better classification results. In order to achieve the above goals, this paper has carried out the following work. This paper studies two mainstream pre-training models in the field of image classification: one is the CNN network based on residual learning; the other is the Vision Transformer model based on Transformer. And according to the performance comparison of each model in four data sets: MNIST, CIFAR-10, CIFAR-100 and ImageNet under different parameters, the optimal model is selected as the background training model of the system. The experimental results show that the Transformer-based model Vision Transformer has better performance and can be used as the back-end training model of the autonomous learning platform.","PeriodicalId":318264,"journal":{"name":"2022 2nd International Conference on Networking Systems of AI (INSAI)","volume":"62 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Research and Implementation of Autonomous Learning Platform for Image Classification\",\"authors\":\"Shiyu Zhao, Menghua Jiang, Zengwen Li, Changxue Chen, Liang Song\",\"doi\":\"10.1109/INSAI56792.2022.00039\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The image classification task aims to automatically classify image content based on machine learning methods. This task is a basic task in the field of computer vision, which has broad application prospects and great research value. At present, in the case of large-scale corpus annotation, mainstream image classification algorithms based on deep learning have been able to obtain better classification results. In order to achieve the above goals, this paper has carried out the following work. This paper studies two mainstream pre-training models in the field of image classification: one is the CNN network based on residual learning; the other is the Vision Transformer model based on Transformer. And according to the performance comparison of each model in four data sets: MNIST, CIFAR-10, CIFAR-100 and ImageNet under different parameters, the optimal model is selected as the background training model of the system. The experimental results show that the Transformer-based model Vision Transformer has better performance and can be used as the back-end training model of the autonomous learning platform.\",\"PeriodicalId\":318264,\"journal\":{\"name\":\"2022 2nd International Conference on Networking Systems of AI (INSAI)\",\"volume\":\"62 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 2nd International Conference on Networking Systems of AI (INSAI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/INSAI56792.2022.00039\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 2nd International Conference on Networking Systems of AI (INSAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/INSAI56792.2022.00039","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

图像分类任务旨在基于机器学习方法对图像内容进行自动分类。该任务是计算机视觉领域的一项基础性任务,具有广阔的应用前景和很大的研究价值。目前,在大规模语料库标注的情况下,基于深度学习的主流图像分类算法已经能够获得较好的分类结果。为了实现上述目标,本文进行了以下工作。本文研究了图像分类领域两种主流的预训练模型:一种是基于残差学习的CNN网络;另一个是基于Transformer的Vision Transformer模型。并根据各模型在MNIST、CIFAR-10、CIFAR-100和ImageNet 4个数据集中不同参数下的性能比较,选择最优模型作为系统的背景训练模型。实验结果表明,基于Transformer的模型具有较好的性能,可以作为自主学习平台的后端训练模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Research and Implementation of Autonomous Learning Platform for Image Classification
The image classification task aims to automatically classify image content based on machine learning methods. This task is a basic task in the field of computer vision, which has broad application prospects and great research value. At present, in the case of large-scale corpus annotation, mainstream image classification algorithms based on deep learning have been able to obtain better classification results. In order to achieve the above goals, this paper has carried out the following work. This paper studies two mainstream pre-training models in the field of image classification: one is the CNN network based on residual learning; the other is the Vision Transformer model based on Transformer. And according to the performance comparison of each model in four data sets: MNIST, CIFAR-10, CIFAR-100 and ImageNet under different parameters, the optimal model is selected as the background training model of the system. The experimental results show that the Transformer-based model Vision Transformer has better performance and can be used as the back-end training model of the autonomous learning platform.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信