Comparing the Robustness of Classical and Deep Learning Techniques for Text Classification

Quynh Tran, Krystsina Shpileuskaya, Elaine Zaunseder, Larissa Putzar, S. Blankenburg
{"title":"Comparing the Robustness of Classical and Deep Learning Techniques for Text Classification","authors":"Quynh Tran, Krystsina Shpileuskaya, Elaine Zaunseder, Larissa Putzar, S. Blankenburg","doi":"10.1109/IJCNN55064.2022.9892242","DOIUrl":null,"url":null,"abstract":"Deep learning algorithms achieve exceptional accuracies in various tasks. Despite this success, those models are known to be prone to errors, i.e. low in robustness, due to differences between training and production environment. One might assume that more model complexity translates directly to more robustness. Therefore, we compare simple, classical models (logistic regression, support vector machine) with complex deep learning techniques (convolutional neural networks, transformers) to provide novel insights into the robustness of machine learning systems. In our approach, we assess the robustness by developing and applying three realistic perturbations, mimicking scanning, typing, and speech recognition errors occurring in inputs for text classification tasks. Hence, we performed a thorough study analyzing the impact of different perturbations with variable strengths on character and word level. A noteworthy finding is that algorithms with low complexity can achieve high robustness. Additionally, we demonstrate that augmented training regarding a specific perturbation can strengthen the chosen models' robustness against other perturbations without reducing their accuracy. Our results can impact the selection of machine learning models and provide a guideline on how to examine the robustness of text classification methods for real-world applications. Moreover, our implementation is publicly available, which contributes to the development of more robust machine learning systems.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Joint Conference on Neural Networks (IJCNN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN55064.2022.9892242","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

Deep learning algorithms achieve exceptional accuracies in various tasks. Despite this success, those models are known to be prone to errors, i.e. low in robustness, due to differences between training and production environment. One might assume that more model complexity translates directly to more robustness. Therefore, we compare simple, classical models (logistic regression, support vector machine) with complex deep learning techniques (convolutional neural networks, transformers) to provide novel insights into the robustness of machine learning systems. In our approach, we assess the robustness by developing and applying three realistic perturbations, mimicking scanning, typing, and speech recognition errors occurring in inputs for text classification tasks. Hence, we performed a thorough study analyzing the impact of different perturbations with variable strengths on character and word level. A noteworthy finding is that algorithms with low complexity can achieve high robustness. Additionally, we demonstrate that augmented training regarding a specific perturbation can strengthen the chosen models' robustness against other perturbations without reducing their accuracy. Our results can impact the selection of machine learning models and provide a guideline on how to examine the robustness of text classification methods for real-world applications. Moreover, our implementation is publicly available, which contributes to the development of more robust machine learning systems.
比较经典和深度学习技术在文本分类中的鲁棒性
深度学习算法在各种任务中实现了卓越的准确性。尽管取得了成功,但由于训练和生产环境之间的差异,这些模型容易出错,即鲁棒性较低。有人可能会认为,更多的模型复杂性直接转化为更强的鲁棒性。因此,我们将简单的经典模型(逻辑回归、支持向量机)与复杂的深度学习技术(卷积神经网络、变压器)进行比较,以提供对机器学习系统鲁棒性的新见解。在我们的方法中,我们通过开发和应用三种现实的扰动来评估鲁棒性,模拟文本分类任务输入中出现的扫描、打字和语音识别错误。因此,我们进行了深入的研究,分析了不同强度的扰动对字符和单词水平的影响。一个值得注意的发现是,低复杂度的算法可以获得高鲁棒性。此外,我们证明了关于特定扰动的增强训练可以增强所选模型对其他扰动的鲁棒性,而不会降低其准确性。我们的研究结果可以影响机器学习模型的选择,并为如何检查现实世界应用中文本分类方法的鲁棒性提供指导。此外,我们的实现是公开的,这有助于开发更强大的机器学习系统。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信