基于空间变换的任意方向喷墨字符识别

Wentao Cai, Hao Zhao, Heng Wang, Xue Deng
{"title":"基于空间变换的任意方向喷墨字符识别","authors":"Wentao Cai, Hao Zhao, Heng Wang, Xue Deng","doi":"10.1109/CCISP55629.2022.9974507","DOIUrl":null,"url":null,"abstract":"Aiming at the problem of the low recognition accuracy caused by the arbitrary characters. In this paper, we propose an arbitrary direction character recognition network. Firstly, a lightweight spatial transformation network (STNet) is designed based on the MobileNetV2, which is used to extract the spatial features of the arbitrary characters and perform spatial transformation. Simultaneously, we introduced the SE attention block into the feature extraction backbone network, which makes the network focuses on the key regions of characters. Then, we build a text recognizer based on recurrent neural network and introduce the Connectionist Temporal Classification (CTC) loss to achieve the flexible alignment between the visual features and the prediction outputs. Extensive experiments are carried out on the IIIT5K and a self-made inkjet characters dataset. The recognition accuracy of our proposed method reaches 95.7% and 86.3% respectively. Compared with the benchmarks, the maximum accuracy of the proposed method is improved by 17.5%. Experimental results show the effectiveness of our proposed method.","PeriodicalId":431851,"journal":{"name":"2022 7th International Conference on Communication, Image and Signal Processing (CCISP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Arbitrary Direction Inkjet Character Recognition Based on Spatial Transformation\",\"authors\":\"Wentao Cai, Hao Zhao, Heng Wang, Xue Deng\",\"doi\":\"10.1109/CCISP55629.2022.9974507\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Aiming at the problem of the low recognition accuracy caused by the arbitrary characters. In this paper, we propose an arbitrary direction character recognition network. Firstly, a lightweight spatial transformation network (STNet) is designed based on the MobileNetV2, which is used to extract the spatial features of the arbitrary characters and perform spatial transformation. Simultaneously, we introduced the SE attention block into the feature extraction backbone network, which makes the network focuses on the key regions of characters. Then, we build a text recognizer based on recurrent neural network and introduce the Connectionist Temporal Classification (CTC) loss to achieve the flexible alignment between the visual features and the prediction outputs. Extensive experiments are carried out on the IIIT5K and a self-made inkjet characters dataset. The recognition accuracy of our proposed method reaches 95.7% and 86.3% respectively. Compared with the benchmarks, the maximum accuracy of the proposed method is improved by 17.5%. Experimental results show the effectiveness of our proposed method.\",\"PeriodicalId\":431851,\"journal\":{\"name\":\"2022 7th International Conference on Communication, Image and Signal Processing (CCISP)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 7th International Conference on Communication, Image and Signal Processing (CCISP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CCISP55629.2022.9974507\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 7th International Conference on Communication, Image and Signal Processing (CCISP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCISP55629.2022.9974507","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

针对任意字符导致识别精度低的问题。本文提出了一种任意方向字符识别网络。首先,基于MobileNetV2设计了一个轻量级空间变换网络(STNet),用于提取任意字符的空间特征并进行空间变换;同时,我们将SE注意块引入特征提取骨干网络,使网络集中在字符的关键区域。然后,我们建立了一个基于递归神经网络的文本识别器,并引入了连接时间分类(CTC)损失来实现视觉特征与预测输出之间的灵活对齐。在IIIT5K和自制的喷墨字符数据集上进行了大量的实验。该方法的识别准确率分别达到95.7%和86.3%。与基准算法相比,该方法的最大准确率提高了17.5%。实验结果表明了该方法的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Arbitrary Direction Inkjet Character Recognition Based on Spatial Transformation
Aiming at the problem of the low recognition accuracy caused by the arbitrary characters. In this paper, we propose an arbitrary direction character recognition network. Firstly, a lightweight spatial transformation network (STNet) is designed based on the MobileNetV2, which is used to extract the spatial features of the arbitrary characters and perform spatial transformation. Simultaneously, we introduced the SE attention block into the feature extraction backbone network, which makes the network focuses on the key regions of characters. Then, we build a text recognizer based on recurrent neural network and introduce the Connectionist Temporal Classification (CTC) loss to achieve the flexible alignment between the visual features and the prediction outputs. Extensive experiments are carried out on the IIIT5K and a self-made inkjet characters dataset. The recognition accuracy of our proposed method reaches 95.7% and 86.3% respectively. Compared with the benchmarks, the maximum accuracy of the proposed method is improved by 17.5%. Experimental results show the effectiveness of our proposed method.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信