Neural Network Verification using Residual Reasoning

Y. Elboher, Elazar Cohen, Guy Katz
{"title":"Neural Network Verification using Residual Reasoning","authors":"Y. Elboher, Elazar Cohen, Guy Katz","doi":"10.48550/arXiv.2208.03083","DOIUrl":null,"url":null,"abstract":"With the increasing integration of neural networks as components in mission-critical systems, there is an increasing need to ensure that they satisfy various safety and liveness requirements. In recent years, numerous sound and complete verification methods have been proposed towards that end, but these typically suffer from severe scalability limitations. Recent work has proposed enhancing such verification techniques with abstraction-refinement capabilities, which have been shown to boost scalability: instead of verifying a large and complex network, the verifier constructs and then verifies a much smaller network, whose correctness implies the correctness of the original network. A shortcoming of such a scheme is that if verifying the smaller network fails, the verifier needs to perform a refinement step that increases the size of the network being verified, and then start verifying the new network from scratch - effectively\"wasting\"its earlier work on verifying the smaller network. In this paper, we present an enhancement to abstraction-based verification of neural networks, by using residual reasoning: the process of utilizing information acquired when verifying an abstract network, in order to expedite the verification of a refined network. In essence, the method allows the verifier to store information about parts of the search space in which the refined network is guaranteed to behave correctly, and allows it to focus on areas where bugs might be discovered. We implemented our approach as an extension to the Marabou verifier, and obtained promising results.","PeriodicalId":413471,"journal":{"name":"IEEE International Conference on Software Engineering and Formal Methods","volume":"174 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-08-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"9","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE International Conference on Software Engineering and Formal Methods","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2208.03083","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 9

Abstract

With the increasing integration of neural networks as components in mission-critical systems, there is an increasing need to ensure that they satisfy various safety and liveness requirements. In recent years, numerous sound and complete verification methods have been proposed towards that end, but these typically suffer from severe scalability limitations. Recent work has proposed enhancing such verification techniques with abstraction-refinement capabilities, which have been shown to boost scalability: instead of verifying a large and complex network, the verifier constructs and then verifies a much smaller network, whose correctness implies the correctness of the original network. A shortcoming of such a scheme is that if verifying the smaller network fails, the verifier needs to perform a refinement step that increases the size of the network being verified, and then start verifying the new network from scratch - effectively"wasting"its earlier work on verifying the smaller network. In this paper, we present an enhancement to abstraction-based verification of neural networks, by using residual reasoning: the process of utilizing information acquired when verifying an abstract network, in order to expedite the verification of a refined network. In essence, the method allows the verifier to store information about parts of the search space in which the refined network is guaranteed to behave correctly, and allows it to focus on areas where bugs might be discovered. We implemented our approach as an extension to the Marabou verifier, and obtained promising results.
残差推理的神经网络验证
随着神经网络作为关键任务系统的组成部分越来越多地集成在一起,人们越来越需要确保神经网络满足各种安全性和活动性要求。近年来,为此提出了许多健全和完整的验证方法,但这些方法通常受到严重的可伸缩性限制。最近的工作提出了用抽象细化功能来增强这种验证技术,这已被证明可以提高可扩展性:验证者不是验证大型复杂的网络,而是构建然后验证一个小得多的网络,其正确性意味着原始网络的正确性。这种方案的一个缺点是,如果验证较小的网络失败,验证者需要执行一个细化步骤,增加被验证网络的规模,然后从头开始验证新网络——有效地“浪费”了其先前验证较小网络的工作。在本文中,我们提出了一种基于抽象的神经网络验证的增强,通过使用残差推理:利用验证抽象网络时获得的信息的过程,以加快对精炼网络的验证。从本质上讲,该方法允许验证者存储有关搜索空间部分的信息,在这些搜索空间中,精炼的网络可以保证正确运行,并允许它专注于可能发现错误的区域。我们将我们的方法作为Marabou验证器的扩展来实现,并获得了令人满意的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信