解码器变压器网络破解低资源实时文本验证码系统的有效性

Rajat Subhra Bhowmick, Isha Ganguli, Jayanta Paul, J. Sil
{"title":"解码器变压器网络破解低资源实时文本验证码系统的有效性","authors":"Rajat Subhra Bhowmick, Isha Ganguli, Jayanta Paul, J. Sil","doi":"10.1109/CW52790.2021.00054","DOIUrl":null,"url":null,"abstract":"Text-based Captchas are still the most extensively used captcha systems in both business and government institutions, notwithstanding its shortcomings. The more easily a captcha system is automatically solved, the greater is the risk to the website. State-of-the-art pretrained hybrid CNN models (combination of CNN with RNN or CNN with Transformer) have been used with large training datasets for image to text (character) sequence applications such as OCR and text captcha solving. However, it is critical to examine the architectures with low resource settings (i.e. minimal weight parameters) so they can be quickly trained for a fresh dataset using limited hardware resources. Due to the low resource context, the models may be easily adapted and trained from scratch for testing the effectiveness of multiple text-based captcha systems. In our study, we focus on the capability of a single encoder based Transformer network to solving a real-time captcha system in low-resource settings. Here we use manually annotated captcha training samples, small in size. Experimental results exhibit that the non-CNN based Transformer approach outperforms the CNN approach even with fewer model weight parameters. For this study, we concentrate on the real-time captcha systems utilized by five Indian government websites. We train the models from scratch using the corresponding manually annotated real-time datasets to demonstrate the vulnerability of each of the captcha system and to report the performance of the models in low resource circumstances. We observe that, though the importance of the information available on these government websites is enormous, the effort required to solve the captcha systems is modest, indicating a potential risk.","PeriodicalId":199618,"journal":{"name":"2021 International Conference on Cyberworlds (CW)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Effectiveness of Decoder Transformer Network in breaking Low-resource Real-time text Captcha System\",\"authors\":\"Rajat Subhra Bhowmick, Isha Ganguli, Jayanta Paul, J. Sil\",\"doi\":\"10.1109/CW52790.2021.00054\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Text-based Captchas are still the most extensively used captcha systems in both business and government institutions, notwithstanding its shortcomings. The more easily a captcha system is automatically solved, the greater is the risk to the website. State-of-the-art pretrained hybrid CNN models (combination of CNN with RNN or CNN with Transformer) have been used with large training datasets for image to text (character) sequence applications such as OCR and text captcha solving. However, it is critical to examine the architectures with low resource settings (i.e. minimal weight parameters) so they can be quickly trained for a fresh dataset using limited hardware resources. Due to the low resource context, the models may be easily adapted and trained from scratch for testing the effectiveness of multiple text-based captcha systems. In our study, we focus on the capability of a single encoder based Transformer network to solving a real-time captcha system in low-resource settings. Here we use manually annotated captcha training samples, small in size. Experimental results exhibit that the non-CNN based Transformer approach outperforms the CNN approach even with fewer model weight parameters. For this study, we concentrate on the real-time captcha systems utilized by five Indian government websites. We train the models from scratch using the corresponding manually annotated real-time datasets to demonstrate the vulnerability of each of the captcha system and to report the performance of the models in low resource circumstances. We observe that, though the importance of the information available on these government websites is enormous, the effort required to solve the captcha systems is modest, indicating a potential risk.\",\"PeriodicalId\":199618,\"journal\":{\"name\":\"2021 International Conference on Cyberworlds (CW)\",\"volume\":\"18 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Conference on Cyberworlds (CW)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CW52790.2021.00054\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Cyberworlds (CW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CW52790.2021.00054","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

基于文本的验证码仍然是商业和政府机构中使用最广泛的验证码系统,尽管它有缺点。验证码系统越容易被自动破解,网站面临的风险就越大。最先进的预训练混合CNN模型(CNN与RNN或CNN与Transformer的组合)已用于大型训练数据集,用于图像到文本(字符)序列应用,如OCR和文本验证码解决。然而,检查具有低资源设置(即最小权重参数)的架构是至关重要的,这样它们可以使用有限的硬件资源快速训练新数据集。由于低资源上下文,模型可以很容易地适应和从头开始训练,以测试多个基于文本的验证码系统的有效性。在我们的研究中,我们专注于基于变压器网络的单个编码器的能力,以解决低资源设置下的实时验证码系统。这里我们使用人工标注的captcha训练样本,规模很小。实验结果表明,即使模型权重参数较少,非基于CNN的Transformer方法也优于CNN方法。在这项研究中,我们将重点放在五个印度政府网站使用的实时验证码系统上。我们使用相应的手动注释的实时数据集从头开始训练模型,以展示每个captcha系统的漏洞,并报告模型在低资源环境下的性能。我们注意到,尽管这些政府网站上提供的信息非常重要,但解决captcha系统所需的努力并不大,这表明存在潜在风险。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Effectiveness of Decoder Transformer Network in breaking Low-resource Real-time text Captcha System
Text-based Captchas are still the most extensively used captcha systems in both business and government institutions, notwithstanding its shortcomings. The more easily a captcha system is automatically solved, the greater is the risk to the website. State-of-the-art pretrained hybrid CNN models (combination of CNN with RNN or CNN with Transformer) have been used with large training datasets for image to text (character) sequence applications such as OCR and text captcha solving. However, it is critical to examine the architectures with low resource settings (i.e. minimal weight parameters) so they can be quickly trained for a fresh dataset using limited hardware resources. Due to the low resource context, the models may be easily adapted and trained from scratch for testing the effectiveness of multiple text-based captcha systems. In our study, we focus on the capability of a single encoder based Transformer network to solving a real-time captcha system in low-resource settings. Here we use manually annotated captcha training samples, small in size. Experimental results exhibit that the non-CNN based Transformer approach outperforms the CNN approach even with fewer model weight parameters. For this study, we concentrate on the real-time captcha systems utilized by five Indian government websites. We train the models from scratch using the corresponding manually annotated real-time datasets to demonstrate the vulnerability of each of the captcha system and to report the performance of the models in low resource circumstances. We observe that, though the importance of the information available on these government websites is enormous, the effort required to solve the captcha systems is modest, indicating a potential risk.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信