{"title":"深度神经网络贝叶斯不确定性估计的分层近似推理","authors":"Ni Zhang, Xiaoyi Chen, Li Quan","doi":"10.1109/IJCNN52387.2021.9534229","DOIUrl":null,"url":null,"abstract":"A proper representation of predictive uncertainty is vital for deep neural networks (DNNs) to be applied in safety-critical domains such as medical diagnosis and self-driving. State-of-the-art (SOTA) variational inference approximation techniques provide a theoretical framework for modeling uncertainty, however, they have not been proven to work on large and deep networks with practical computation. In this study, we develop a layerwise approximation with a local reparameterization technique to efficiently perform sophisticated variational Bayesian inference on very deep SOTA convolutional neural networks (CNNs) (VGG16, ResNet variants, DenseNet). Theoretical analysis is presented to justify that the layerwise approach remains a Bayesian neural network. We further derive a SOTA $\\alpha$-divergence objective function to work with the layerwise approximate inference, addressing the concern of underestimating uncertainties by the Kullback-Leibler divergence. Empirical evaluation using MNIST, CIFAR-10, and CIFAR-100 datasets consistently shows that with our proposal, deep CNN models can have a better quality of predictive uncertainty than Monte Carlo-dropout in detecting in-domain misclassification and excel in out-of-distribution detection.","PeriodicalId":396583,"journal":{"name":"2021 International Joint Conference on Neural Networks (IJCNN)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Layerwise Approximate Inference for Bayesian Uncertainty Estimates on Deep Neural Networks\",\"authors\":\"Ni Zhang, Xiaoyi Chen, Li Quan\",\"doi\":\"10.1109/IJCNN52387.2021.9534229\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A proper representation of predictive uncertainty is vital for deep neural networks (DNNs) to be applied in safety-critical domains such as medical diagnosis and self-driving. State-of-the-art (SOTA) variational inference approximation techniques provide a theoretical framework for modeling uncertainty, however, they have not been proven to work on large and deep networks with practical computation. In this study, we develop a layerwise approximation with a local reparameterization technique to efficiently perform sophisticated variational Bayesian inference on very deep SOTA convolutional neural networks (CNNs) (VGG16, ResNet variants, DenseNet). Theoretical analysis is presented to justify that the layerwise approach remains a Bayesian neural network. We further derive a SOTA $\\\\alpha$-divergence objective function to work with the layerwise approximate inference, addressing the concern of underestimating uncertainties by the Kullback-Leibler divergence. Empirical evaluation using MNIST, CIFAR-10, and CIFAR-100 datasets consistently shows that with our proposal, deep CNN models can have a better quality of predictive uncertainty than Monte Carlo-dropout in detecting in-domain misclassification and excel in out-of-distribution detection.\",\"PeriodicalId\":396583,\"journal\":{\"name\":\"2021 International Joint Conference on Neural Networks (IJCNN)\",\"volume\":\"7 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Joint Conference on Neural Networks (IJCNN)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IJCNN52387.2021.9534229\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Joint Conference on Neural Networks (IJCNN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN52387.2021.9534229","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Layerwise Approximate Inference for Bayesian Uncertainty Estimates on Deep Neural Networks
A proper representation of predictive uncertainty is vital for deep neural networks (DNNs) to be applied in safety-critical domains such as medical diagnosis and self-driving. State-of-the-art (SOTA) variational inference approximation techniques provide a theoretical framework for modeling uncertainty, however, they have not been proven to work on large and deep networks with practical computation. In this study, we develop a layerwise approximation with a local reparameterization technique to efficiently perform sophisticated variational Bayesian inference on very deep SOTA convolutional neural networks (CNNs) (VGG16, ResNet variants, DenseNet). Theoretical analysis is presented to justify that the layerwise approach remains a Bayesian neural network. We further derive a SOTA $\alpha$-divergence objective function to work with the layerwise approximate inference, addressing the concern of underestimating uncertainties by the Kullback-Leibler divergence. Empirical evaluation using MNIST, CIFAR-10, and CIFAR-100 datasets consistently shows that with our proposal, deep CNN models can have a better quality of predictive uncertainty than Monte Carlo-dropout in detecting in-domain misclassification and excel in out-of-distribution detection.