Robustness Analysis of Neural Network Designs for ReLU Family and Batch Normalization

Hang Chen, Yi-Pei Su, Yean-Ru Chen, Chi-Chieh Chiu, Sao-Jie Chen
{"title":"Robustness Analysis of Neural Network Designs for ReLU Family and Batch Normalization","authors":"Hang Chen, Yi-Pei Su, Yean-Ru Chen, Chi-Chieh Chiu, Sao-Jie Chen","doi":"10.1109/TAAI57707.2022.00010","DOIUrl":null,"url":null,"abstract":"The so-called neural network (NN) robustness problem, its original definition is that if an image is perturbed, the classification result of the image can still maintain the original correct category. It means that at the end of the entire NN operation, the lower bound of the target category value must be greater than the upper bound value of all other categories. There are multiple design techniques which can either bring the neural networks higher accuracy or maintain accuracy while reducing the computation effort at the same time. However, very few work focus on giving an efficient and reliable estimation of the trend of robustness changing directly with respect to the design factors. Lacking such information would either damage on designing a robust NN for critical systems or postpone the robustness analysis after the design completed and then results in paying more cost on NN design modifications. In this paper, we not only provide numerous experimental results but also propose three extended lemmas based on the related work which analyzes robustness with Lipschitz constant, to discuss how the two commonly used design factors, the ReLU based activation functions and batch normalization technique, bring the effectiveness to the robustness changing trend, under the condition of that the NN can still retain acceptable accuracy. We can conclude that we encourage to adopt ReLU than its other family activation functions (e.g. Leaky-ReLU and ELU) but discourage to use batch normalization compared with adopting it in the same NN design if we expect for higher robustness.","PeriodicalId":111620,"journal":{"name":"2022 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/TAAI57707.2022.00010","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The so-called neural network (NN) robustness problem, its original definition is that if an image is perturbed, the classification result of the image can still maintain the original correct category. It means that at the end of the entire NN operation, the lower bound of the target category value must be greater than the upper bound value of all other categories. There are multiple design techniques which can either bring the neural networks higher accuracy or maintain accuracy while reducing the computation effort at the same time. However, very few work focus on giving an efficient and reliable estimation of the trend of robustness changing directly with respect to the design factors. Lacking such information would either damage on designing a robust NN for critical systems or postpone the robustness analysis after the design completed and then results in paying more cost on NN design modifications. In this paper, we not only provide numerous experimental results but also propose three extended lemmas based on the related work which analyzes robustness with Lipschitz constant, to discuss how the two commonly used design factors, the ReLU based activation functions and batch normalization technique, bring the effectiveness to the robustness changing trend, under the condition of that the NN can still retain acceptable accuracy. We can conclude that we encourage to adopt ReLU than its other family activation functions (e.g. Leaky-ReLU and ELU) but discourage to use batch normalization compared with adopting it in the same NN design if we expect for higher robustness.
ReLU族与批归一化神经网络设计的鲁棒性分析
所谓神经网络(NN)的鲁棒性问题,其最初的定义是,如果图像受到扰动,图像的分类结果仍能保持原来正确的类别。这意味着在整个NN操作结束时,目标类别值的下界必须大于所有其他类别值的上界。有多种设计技术既可以提高神经网络的精度,又可以在保持精度的同时减少计算量。然而,很少有工作关注于对鲁棒性直接随设计因素变化的趋势给出有效和可靠的估计。如果缺乏这些信息,要么会破坏关键系统鲁棒神经网络的设计,要么会推迟设计完成后的鲁棒性分析,从而导致在神经网络设计修改上付出更多的成本。在本文中,我们不仅提供了大量的实验结果,而且在Lipschitz常数分析鲁棒性的相关工作的基础上,提出了三个扩展引理,讨论了两种常用的设计因素,即基于ReLU的激活函数和批归一化技术,如何在神经网络仍能保持可接受精度的情况下,对鲁棒性变化趋势的有效性。我们可以得出结论,我们鼓励采用ReLU而不是其他家族激活函数(例如Leaky-ReLU和ELU),但如果我们期望更高的鲁棒性,则不鼓励使用批归一化,而不是在相同的NN设计中采用它。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信