Jordan F. Masakuna;Djeff K. Nkashama;Arian Soltani;Marc Frappier;Pierre M. Tardif;Froduald Kabanza
{"title":"Streamlined and Resource-Efficient Estimation of Epistemic Uncertainty in Deep Ensemble Classification Decision via Regression","authors":"Jordan F. Masakuna;Djeff K. Nkashama;Arian Soltani;Marc Frappier;Pierre M. Tardif;Froduald Kabanza","doi":"10.1109/TETCI.2024.3508846","DOIUrl":null,"url":null,"abstract":"Ensemble deep learning (EDL) has emerged as a leading tool for epistemic uncertainty quantification (UQ) in predictive modelling. Our study focuses on the utilization of EDL, composed of auto-encoders (AEs) for out-of-distribution (OoD) detection. EDL offers straightforward interpretability and valuable practical insights. Conventionally, employing multiple AEs in an ensemble requires regular training for each model whenever substantial changes occur in the data, a process that can become computationally expensive, especially when dealing with large ensembles. To address this computational challenge, we introduce an innovative strategy that treats ensemble UQ as a regression problem. During initial training, once the uncertainty distribution is established, we map this distribution to one ensemble member. This approach ensures that during subsequent trainings and inferences, only one ensemble member and the regression model are needed to predict uncertainties, eliminating the need to maintain the entire ensemble. This streamlined approach is particularly advantageous for systems with limited computational resources or situations that demand rapid decision-making, such as alert management in cybersecurity. Our evaluations on five benchmark OoD detection data sets demonstrate that the uncertainty estimates obtained with our proposed method can, in most cases, align with the uncertainty distribution learned by the ensemble, all while significantly reducing the computational resource requirements.","PeriodicalId":13135,"journal":{"name":"IEEE Transactions on Emerging Topics in Computational Intelligence","volume":"9 4","pages":"2940-2951"},"PeriodicalIF":5.3000,"publicationDate":"2024-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Emerging Topics in Computational Intelligence","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10787107/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Ensemble deep learning (EDL) has emerged as a leading tool for epistemic uncertainty quantification (UQ) in predictive modelling. Our study focuses on the utilization of EDL, composed of auto-encoders (AEs) for out-of-distribution (OoD) detection. EDL offers straightforward interpretability and valuable practical insights. Conventionally, employing multiple AEs in an ensemble requires regular training for each model whenever substantial changes occur in the data, a process that can become computationally expensive, especially when dealing with large ensembles. To address this computational challenge, we introduce an innovative strategy that treats ensemble UQ as a regression problem. During initial training, once the uncertainty distribution is established, we map this distribution to one ensemble member. This approach ensures that during subsequent trainings and inferences, only one ensemble member and the regression model are needed to predict uncertainties, eliminating the need to maintain the entire ensemble. This streamlined approach is particularly advantageous for systems with limited computational resources or situations that demand rapid decision-making, such as alert management in cybersecurity. Our evaluations on five benchmark OoD detection data sets demonstrate that the uncertainty estimates obtained with our proposed method can, in most cases, align with the uncertainty distribution learned by the ensemble, all while significantly reducing the computational resource requirements.
集成深度学习(EDL)已成为预测建模中认知不确定性量化(UQ)的主要工具。我们的研究重点是利用由自编码器(ae)组成的EDL进行out- distribution (OoD)检测。EDL提供了直接的可解释性和有价值的实用见解。通常,在集成中使用多个ae需要在数据发生重大变化时对每个模型进行定期训练,这个过程可能会变得计算昂贵,特别是在处理大型集成时。为了解决这一计算挑战,我们引入了一种创新的策略,将集成UQ视为回归问题。在初始训练过程中,一旦不确定性分布建立,我们将该分布映射到一个集成成员。这种方法确保在随后的训练和推理中,只需要一个集成成员和回归模型来预测不确定性,从而消除了维护整个集成的需要。这种简化的方法对于计算资源有限的系统或需要快速决策的情况特别有利,例如网络安全中的警报管理。我们对五个基准OoD检测数据集的评估表明,在大多数情况下,使用我们提出的方法获得的不确定性估计可以与集成学习的不确定性分布一致,同时显着减少了计算资源需求。
期刊介绍:
The IEEE Transactions on Emerging Topics in Computational Intelligence (TETCI) publishes original articles on emerging aspects of computational intelligence, including theory, applications, and surveys.
TETCI is an electronics only publication. TETCI publishes six issues per year.
Authors are encouraged to submit manuscripts in any emerging topic in computational intelligence, especially nature-inspired computing topics not covered by other IEEE Computational Intelligence Society journals. A few such illustrative examples are glial cell networks, computational neuroscience, Brain Computer Interface, ambient intelligence, non-fuzzy computing with words, artificial life, cultural learning, artificial endocrine networks, social reasoning, artificial hormone networks, computational intelligence for the IoT and Smart-X technologies.