{"title":"FAST:通过特征选择提升基于不确定性的神经网络测试优先级方法","authors":"Jialuo Chen, Jingyi Wang, Xiyue Zhang, Youcheng Sun, Marta Kwiatkowska, Jiming Chen, Peng Cheng","doi":"arxiv-2409.09130","DOIUrl":null,"url":null,"abstract":"Due to the vast testing space, the increasing demand for effective and\nefficient testing of deep neural networks (DNNs) has led to the development of\nvarious DNN test case prioritization techniques. However, the fact that DNNs\ncan deliver high-confidence predictions for incorrectly predicted examples,\nknown as the over-confidence problem, causes these methods to fail to reveal\nhigh-confidence errors. To address this limitation, in this work, we propose\nFAST, a method that boosts existing prioritization methods through guided\nFeAture SelecTion. FAST is based on the insight that certain features may\nintroduce noise that affects the model's output confidence, thereby\ncontributing to high-confidence errors. It quantifies the importance of each\nfeature for the model's correct predictions, and then dynamically prunes the\ninformation from the noisy features during inference to derive a new\nprobability vector for the uncertainty estimation. With the help of FAST, the\nhigh-confidence errors and correctly classified examples become more\ndistinguishable, resulting in higher APFD (Average Percentage of Fault\nDetection) values for test prioritization, and higher generalization ability\nfor model enhancement. We conduct extensive experiments to evaluate FAST across\na diverse set of model structures on multiple benchmark datasets to validate\nthe effectiveness, efficiency, and scalability of FAST compared to the\nstate-of-the-art prioritization techniques.","PeriodicalId":501278,"journal":{"name":"arXiv - CS - Software Engineering","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"FAST: Boosting Uncertainty-based Test Prioritization Methods for Neural Networks via Feature Selection\",\"authors\":\"Jialuo Chen, Jingyi Wang, Xiyue Zhang, Youcheng Sun, Marta Kwiatkowska, Jiming Chen, Peng Cheng\",\"doi\":\"arxiv-2409.09130\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Due to the vast testing space, the increasing demand for effective and\\nefficient testing of deep neural networks (DNNs) has led to the development of\\nvarious DNN test case prioritization techniques. However, the fact that DNNs\\ncan deliver high-confidence predictions for incorrectly predicted examples,\\nknown as the over-confidence problem, causes these methods to fail to reveal\\nhigh-confidence errors. To address this limitation, in this work, we propose\\nFAST, a method that boosts existing prioritization methods through guided\\nFeAture SelecTion. FAST is based on the insight that certain features may\\nintroduce noise that affects the model's output confidence, thereby\\ncontributing to high-confidence errors. It quantifies the importance of each\\nfeature for the model's correct predictions, and then dynamically prunes the\\ninformation from the noisy features during inference to derive a new\\nprobability vector for the uncertainty estimation. With the help of FAST, the\\nhigh-confidence errors and correctly classified examples become more\\ndistinguishable, resulting in higher APFD (Average Percentage of Fault\\nDetection) values for test prioritization, and higher generalization ability\\nfor model enhancement. We conduct extensive experiments to evaluate FAST across\\na diverse set of model structures on multiple benchmark datasets to validate\\nthe effectiveness, efficiency, and scalability of FAST compared to the\\nstate-of-the-art prioritization techniques.\",\"PeriodicalId\":501278,\"journal\":{\"name\":\"arXiv - CS - Software Engineering\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Software Engineering\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.09130\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Software Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09130","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
由于测试空间广阔,对深度神经网络(DNN)进行有效和高效测试的需求与日俱增,因此开发了各种 DNN 测试用例优先级排序技术。然而,由于 DNN 可以为预测错误的示例提供高置信度预测,即所谓的过置信度问题,导致这些方法无法揭示高置信度错误。为了解决这一局限性,我们在这项工作中提出了 FAST 方法,该方法通过引导图像选择来提升现有的优先级排序方法。FAST 基于这样一种认识:某些特征可能会带来噪声,影响模型输出的置信度,从而导致高置信度错误。它量化了每个特征对模型正确预测的重要性,然后在推理过程中动态修剪噪声特征信息,为不确定性估计导出新的概率向量。在 FAST 的帮助下,高置信度错误和正确分类的示例变得更加难以区分,从而为测试优先级的确定带来更高的 APFD(平均故障检测百分比)值,并为模型的增强带来更高的泛化能力。我们进行了广泛的实验,在多个基准数据集上对 FAST 的各种模型结构进行了评估,从而验证了 FAST 与最先进的优先级排序技术相比的有效性、效率和可扩展性。
FAST: Boosting Uncertainty-based Test Prioritization Methods for Neural Networks via Feature Selection
Due to the vast testing space, the increasing demand for effective and
efficient testing of deep neural networks (DNNs) has led to the development of
various DNN test case prioritization techniques. However, the fact that DNNs
can deliver high-confidence predictions for incorrectly predicted examples,
known as the over-confidence problem, causes these methods to fail to reveal
high-confidence errors. To address this limitation, in this work, we propose
FAST, a method that boosts existing prioritization methods through guided
FeAture SelecTion. FAST is based on the insight that certain features may
introduce noise that affects the model's output confidence, thereby
contributing to high-confidence errors. It quantifies the importance of each
feature for the model's correct predictions, and then dynamically prunes the
information from the noisy features during inference to derive a new
probability vector for the uncertainty estimation. With the help of FAST, the
high-confidence errors and correctly classified examples become more
distinguishable, resulting in higher APFD (Average Percentage of Fault
Detection) values for test prioritization, and higher generalization ability
for model enhancement. We conduct extensive experiments to evaluate FAST across
a diverse set of model structures on multiple benchmark datasets to validate
the effectiveness, efficiency, and scalability of FAST compared to the
state-of-the-art prioritization techniques.