Dropout Injection at Test Time for Post Hoc Uncertainty Quantification in Neural Networks

Emanuele Ledda, G. Fumera, F. Roli
{"title":"Dropout Injection at Test Time for Post Hoc Uncertainty Quantification in Neural Networks","authors":"Emanuele Ledda, G. Fumera, F. Roli","doi":"10.48550/arXiv.2302.02924","DOIUrl":null,"url":null,"abstract":"Among Bayesian methods, Monte-Carlo dropout provides principled tools for evaluating the epistemic uncertainty of neural networks. Its popularity recently led to seminal works that proposed activating the dropout layers only during inference for evaluating uncertainty. This approach, which we call dropout injection, provides clear benefits over its traditional counterpart (which we call embedded dropout) since it allows one to obtain a post hoc uncertainty measure for any existing network previously trained without dropout, avoiding an additional, time-consuming training process. Unfortunately, no previous work compared injected and embedded dropout; therefore, we provide the first thorough investigation, focusing on regression problems. The main contribution of our work is to provide guidelines on the effective use of injected dropout so that it can be a practical alternative to the current use of embedded dropout. In particular, we show that its effectiveness strongly relies on a suitable scaling of the corresponding uncertainty measure, and we discuss the trade-off between negative log-likelihood and calibration error as a function of the scale factor. Experimental results on UCI data sets and crowd counting benchmarks support our claim that dropout injection can effectively behave as a competitive post hoc uncertainty quantification technique.","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"1 1","pages":"119356"},"PeriodicalIF":0.0000,"publicationDate":"2023-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Inf. Sci.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2302.02924","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Among Bayesian methods, Monte-Carlo dropout provides principled tools for evaluating the epistemic uncertainty of neural networks. Its popularity recently led to seminal works that proposed activating the dropout layers only during inference for evaluating uncertainty. This approach, which we call dropout injection, provides clear benefits over its traditional counterpart (which we call embedded dropout) since it allows one to obtain a post hoc uncertainty measure for any existing network previously trained without dropout, avoiding an additional, time-consuming training process. Unfortunately, no previous work compared injected and embedded dropout; therefore, we provide the first thorough investigation, focusing on regression problems. The main contribution of our work is to provide guidelines on the effective use of injected dropout so that it can be a practical alternative to the current use of embedded dropout. In particular, we show that its effectiveness strongly relies on a suitable scaling of the corresponding uncertainty measure, and we discuss the trade-off between negative log-likelihood and calibration error as a function of the scale factor. Experimental results on UCI data sets and crowd counting benchmarks support our claim that dropout injection can effectively behave as a competitive post hoc uncertainty quantification technique.
神经网络事后不确定度量化的测试时间Dropout注入
在贝叶斯方法中,蒙特卡罗dropout为评估神经网络的认知不确定性提供了原则性的工具。它的流行最近导致了开创性的工作,建议仅在评估不确定性的推理过程中激活退出层。这种方法,我们称之为dropout注入,与传统的对应方法(我们称之为嵌入式dropout)相比,它提供了明显的好处,因为它允许人们获得任何先前没有dropout训练的现有网络的事后不确定性度量,避免了额外的,耗时的训练过程。不幸的是,之前没有研究对注射型和嵌入型辍学进行比较;因此,我们提供了第一个彻底的调查,重点是回归问题。我们的工作的主要贡献是提供了有效使用注射型dropout的指导方针,使其能够成为目前使用的嵌入式dropout的实际替代方案。特别是,我们表明其有效性强烈依赖于相应不确定度测量的合适尺度,并且我们讨论了负对数似然和校准误差作为尺度因子的函数之间的权衡。在UCI数据集和人群计数基准上的实验结果支持了我们的观点,即辍学注入可以有效地作为一种有竞争力的事后不确定性量化技术。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信