Verifying the Generalization of Deep Learning to Out-of-Distribution Domains

IF 0.9 3区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Guy Amir, Osher Maayan, Tom Zelazny, Guy Katz, Michael Schapira
{"title":"Verifying the Generalization of Deep Learning to Out-of-Distribution Domains","authors":"Guy Amir, Osher Maayan, Tom Zelazny, Guy Katz, Michael Schapira","doi":"10.1007/s10817-024-09704-7","DOIUrl":null,"url":null,"abstract":"<p>Deep neural networks (DNNs) play a crucial role in the field of machine learning, demonstrating state-of-the-art performance across various application domains. However, despite their success, DNN-based models may occasionally exhibit challenges with <i>generalization</i>, i.e., may fail to handle inputs that were not encountered during training. This limitation is a significant challenge when it comes to deploying deep learning for safety-critical tasks, as well as in real-world settings characterized by substantial variability. We introduce a novel approach for harnessing DNN verification technology to identify DNN-driven decision rules that exhibit robust generalization to previously unencountered input domains. Our method assesses generalization within an input domain by measuring the level of agreement between <i>independently trained</i> deep neural networks for inputs in this domain. We also efficiently realize our approach by using off-the-shelf DNN verification engines, and extensively evaluate it on both supervised and unsupervised DNN benchmarks, including a deep reinforcement learning (DRL) system for Internet congestion control—demonstrating the applicability of our approach for real-world settings. Moreover, our research introduces a fresh objective for formal verification, offering the prospect of mitigating the challenges linked to deploying DNN-driven systems in real-world scenarios.</p>","PeriodicalId":15082,"journal":{"name":"Journal of Automated Reasoning","volume":"1 1","pages":""},"PeriodicalIF":0.9000,"publicationDate":"2024-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Automated Reasoning","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s10817-024-09704-7","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Deep neural networks (DNNs) play a crucial role in the field of machine learning, demonstrating state-of-the-art performance across various application domains. However, despite their success, DNN-based models may occasionally exhibit challenges with generalization, i.e., may fail to handle inputs that were not encountered during training. This limitation is a significant challenge when it comes to deploying deep learning for safety-critical tasks, as well as in real-world settings characterized by substantial variability. We introduce a novel approach for harnessing DNN verification technology to identify DNN-driven decision rules that exhibit robust generalization to previously unencountered input domains. Our method assesses generalization within an input domain by measuring the level of agreement between independently trained deep neural networks for inputs in this domain. We also efficiently realize our approach by using off-the-shelf DNN verification engines, and extensively evaluate it on both supervised and unsupervised DNN benchmarks, including a deep reinforcement learning (DRL) system for Internet congestion control—demonstrating the applicability of our approach for real-world settings. Moreover, our research introduces a fresh objective for formal verification, offering the prospect of mitigating the challenges linked to deploying DNN-driven systems in real-world scenarios.

Abstract Image

验证深度学习在分布外领域的通用性
深度神经网络(DNN)在机器学习领域发挥着至关重要的作用,在各种应用领域都表现出最先进的性能。然而,尽管取得了成功,基于 DNN 的模型偶尔也会在泛化方面遇到挑战,即可能无法处理在训练过程中未遇到的输入。在将深度学习应用于安全关键任务时,以及在以巨大变异性为特征的真实世界环境中,这种局限性是一个重大挑战。我们介绍了一种利用 DNN 验证技术的新方法,用于识别 DNN 驱动的决策规则,这些规则对以前未遇到的输入域具有强大的泛化能力。我们的方法通过测量独立训练的深度神经网络对输入域中输入的一致程度来评估输入域内的泛化。我们还利用现成的 DNN 验证引擎高效地实现了我们的方法,并在有监督和无监督 DNN 基准(包括用于互联网拥塞控制的深度强化学习(DRL)系统)上进行了广泛评估,证明了我们的方法在现实世界中的适用性。此外,我们的研究还为形式验证引入了一个全新的目标,为减轻与在现实世界场景中部署 DNN 驱动型系统相关的挑战提供了前景。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Journal of Automated Reasoning
Journal of Automated Reasoning 工程技术-计算机:人工智能
CiteScore
3.60
自引率
9.10%
发文量
31
审稿时长
>12 weeks
期刊介绍: The Journal of Automated Reasoning is an interdisciplinary journal that maintains a balance between theory, implementation and application. The spectrum of material published ranges from the presentation of a new inference rule with proof of its logical properties to a detailed account of a computer program designed to solve various problems in industry. The main fields covered are automated theorem proving, logic programming, expert systems, program synthesis and validation, artificial intelligence, computational logic, robotics, and various industrial applications. The papers share the common feature of focusing on several aspects of automated reasoning, a field whose objective is the design and implementation of a computer program that serves as an assistant in solving problems and in answering questions that require reasoning. The Journal of Automated Reasoning provides a forum and a means for exchanging information for those interested purely in theory, those interested primarily in implementation, and those interested in specific research and industrial applications.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信