Layerwise Approximate Inference for Bayesian Uncertainty Estimates on Deep Neural Networks

Ni Zhang, Xiaoyi Chen, Li Quan
{"title":"Layerwise Approximate Inference for Bayesian Uncertainty Estimates on Deep Neural Networks","authors":"Ni Zhang, Xiaoyi Chen, Li Quan","doi":"10.1109/IJCNN52387.2021.9534229","DOIUrl":null,"url":null,"abstract":"A proper representation of predictive uncertainty is vital for deep neural networks (DNNs) to be applied in safety-critical domains such as medical diagnosis and self-driving. State-of-the-art (SOTA) variational inference approximation techniques provide a theoretical framework for modeling uncertainty, however, they have not been proven to work on large and deep networks with practical computation. In this study, we develop a layerwise approximation with a local reparameterization technique to efficiently perform sophisticated variational Bayesian inference on very deep SOTA convolutional neural networks (CNNs) (VGG16, ResNet variants, DenseNet). Theoretical analysis is presented to justify that the layerwise approach remains a Bayesian neural network. We further derive a SOTA $\\alpha$-divergence objective function to work with the layerwise approximate inference, addressing the concern of underestimating uncertainties by the Kullback-Leibler divergence. Empirical evaluation using MNIST, CIFAR-10, and CIFAR-100 datasets consistently shows that with our proposal, deep CNN models can have a better quality of predictive uncertainty than Monte Carlo-dropout in detecting in-domain misclassification and excel in out-of-distribution detection.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Joint Conference on Neural Networks (IJCNN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN52387.2021.9534229","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

A proper representation of predictive uncertainty is vital for deep neural networks (DNNs) to be applied in safety-critical domains such as medical diagnosis and self-driving. State-of-the-art (SOTA) variational inference approximation techniques provide a theoretical framework for modeling uncertainty, however, they have not been proven to work on large and deep networks with practical computation. In this study, we develop a layerwise approximation with a local reparameterization technique to efficiently perform sophisticated variational Bayesian inference on very deep SOTA convolutional neural networks (CNNs) (VGG16, ResNet variants, DenseNet). Theoretical analysis is presented to justify that the layerwise approach remains a Bayesian neural network. We further derive a SOTA $\alpha$-divergence objective function to work with the layerwise approximate inference, addressing the concern of underestimating uncertainties by the Kullback-Leibler divergence. Empirical evaluation using MNIST, CIFAR-10, and CIFAR-100 datasets consistently shows that with our proposal, deep CNN models can have a better quality of predictive uncertainty than Monte Carlo-dropout in detecting in-domain misclassification and excel in out-of-distribution detection.
深度神经网络贝叶斯不确定性估计的分层近似推理
预测不确定性的适当表示对于深度神经网络(dnn)在医疗诊断和自动驾驶等安全关键领域的应用至关重要。最先进的(SOTA)变分推理近似技术为建模不确定性提供了理论框架,然而,它们尚未被证明可以在具有实际计算的大型深度网络上工作。在本研究中,我们开发了一种具有局部重参数化技术的分层近似,以有效地在非常深的SOTA卷积神经网络(cnn) (VGG16, ResNet变体,DenseNet)上执行复杂的变分贝叶斯推理。理论分析证明了分层方法仍然是贝叶斯神经网络。我们进一步推导了一个SOTA $\alpha$-divergence目标函数来处理分层近似推理,解决了Kullback-Leibler散度低估不确定性的问题。使用MNIST、CIFAR-10和CIFAR-100数据集进行的经验评估一致表明,在我们的建议下,深度CNN模型在检测域内误分类方面比Monte Carlo-dropout具有更好的预测不确定性质量,并且在检测分布外方面表现出色。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信