基于不一致性的主动学习与自适应伪标记用于鱼类物种识别

M. M. Nabi, Chiranjibi Shah, S. Y. Alaba, Ryan Caillouet, Jack H. Prior, Matthew Campbell, Farron Wallace, John E. Ball, Robert J. Moorhead
{"title":"基于不一致性的主动学习与自适应伪标记用于鱼类物种识别","authors":"M. M. Nabi, Chiranjibi Shah, S. Y. Alaba, Ryan Caillouet, Jack H. Prior, Matthew Campbell, Farron Wallace, John E. Ball, Robert J. Moorhead","doi":"10.1117/12.3013336","DOIUrl":null,"url":null,"abstract":"The deep neural network has found widespread application in object detection due to its high accuracy. However, its performance typically depends on the availability of a substantial volume of accurately labeled data. Several active learning approaches have been proposed to reduce the labeling dependency based on the confidence of the detector. Nevertheless, these approaches tend to exhibit biases toward high-performing classes, resulting in datasets that do not adequately represent the testing data. In this study, we introduce a comprehensive framework for active learning that considers both the uncertainty and the robustness of the detector, ensuring superior performance across all classes. The robustness-based score for active learning is calculated using the consistency between an image and its augmented version. Additionally, we leverage pseudo-labeling to mitigate potential distribution drift and enhance model performance. To address the challenge of setting the pseudo-labeling threshold, we introduce an adaptive threshold mechanism. This adaptability is crucial, as a fixed threshold can negatively impact performance, particularly for low-performing classes or during the initial stages of training. For our experiment, we employ the Southeast Area Monitoring and Assessment Program Dataset 2021 (SEAMAPD21), comprising 130 fish species classes with 28,328 image samples. The results show that our model outperforms the state-of-the-art method and significantly reduces the annotation cost. Furthermore, we benchmark our model’s performance against a public dataset (PASCAL VOC07), showcasing its effectiveness in comparison to existing methods.","PeriodicalId":178341,"journal":{"name":"Defense + Commercial Sensing","volume":"33 S122","pages":"130610C - 130610C-12"},"PeriodicalIF":0.0000,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Inconsistency-based active learning with adaptive pseudo-labeling for fish species identification\",\"authors\":\"M. M. Nabi, Chiranjibi Shah, S. Y. Alaba, Ryan Caillouet, Jack H. Prior, Matthew Campbell, Farron Wallace, John E. Ball, Robert J. Moorhead\",\"doi\":\"10.1117/12.3013336\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The deep neural network has found widespread application in object detection due to its high accuracy. However, its performance typically depends on the availability of a substantial volume of accurately labeled data. Several active learning approaches have been proposed to reduce the labeling dependency based on the confidence of the detector. Nevertheless, these approaches tend to exhibit biases toward high-performing classes, resulting in datasets that do not adequately represent the testing data. In this study, we introduce a comprehensive framework for active learning that considers both the uncertainty and the robustness of the detector, ensuring superior performance across all classes. The robustness-based score for active learning is calculated using the consistency between an image and its augmented version. Additionally, we leverage pseudo-labeling to mitigate potential distribution drift and enhance model performance. To address the challenge of setting the pseudo-labeling threshold, we introduce an adaptive threshold mechanism. This adaptability is crucial, as a fixed threshold can negatively impact performance, particularly for low-performing classes or during the initial stages of training. For our experiment, we employ the Southeast Area Monitoring and Assessment Program Dataset 2021 (SEAMAPD21), comprising 130 fish species classes with 28,328 image samples. The results show that our model outperforms the state-of-the-art method and significantly reduces the annotation cost. Furthermore, we benchmark our model’s performance against a public dataset (PASCAL VOC07), showcasing its effectiveness in comparison to existing methods.\",\"PeriodicalId\":178341,\"journal\":{\"name\":\"Defense + Commercial Sensing\",\"volume\":\"33 S122\",\"pages\":\"130610C - 130610C-12\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-06-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Defense + Commercial Sensing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1117/12.3013336\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Defense + Commercial Sensing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.3013336","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

深度神经网络因其高精度而被广泛应用于物体检测。然而,其性能通常取决于是否有大量准确标注的数据。目前已经提出了几种主动学习方法,以减少基于检测器置信度的标记依赖性。然而,这些方法往往会对表现优异的类别产生偏差,导致数据集不能充分代表测试数据。在本研究中,我们引入了一个全面的主动学习框架,该框架同时考虑了检测器的不确定性和鲁棒性,确保在所有类别中都能获得卓越的性能。基于鲁棒性的主动学习得分是通过图像与其增强版本之间的一致性来计算的。此外,我们还利用伪标记来减轻潜在的分布漂移并提高模型性能。为了解决设置伪标记阈值的难题,我们引入了自适应阈值机制。这种适应性至关重要,因为固定的阈值会对性能产生负面影响,尤其是在低性能类或训练的初始阶段。在实验中,我们使用了东南地区监测和评估计划数据集 2021(SEAMAPD21),其中包括 130 个鱼类物种类别和 28328 个图像样本。结果表明,我们的模型优于最先进的方法,并显著降低了标注成本。此外,我们还以一个公共数据集(PASCAL VOC07)为基准测试了我们模型的性能,展示了它与现有方法相比的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Inconsistency-based active learning with adaptive pseudo-labeling for fish species identification
The deep neural network has found widespread application in object detection due to its high accuracy. However, its performance typically depends on the availability of a substantial volume of accurately labeled data. Several active learning approaches have been proposed to reduce the labeling dependency based on the confidence of the detector. Nevertheless, these approaches tend to exhibit biases toward high-performing classes, resulting in datasets that do not adequately represent the testing data. In this study, we introduce a comprehensive framework for active learning that considers both the uncertainty and the robustness of the detector, ensuring superior performance across all classes. The robustness-based score for active learning is calculated using the consistency between an image and its augmented version. Additionally, we leverage pseudo-labeling to mitigate potential distribution drift and enhance model performance. To address the challenge of setting the pseudo-labeling threshold, we introduce an adaptive threshold mechanism. This adaptability is crucial, as a fixed threshold can negatively impact performance, particularly for low-performing classes or during the initial stages of training. For our experiment, we employ the Southeast Area Monitoring and Assessment Program Dataset 2021 (SEAMAPD21), comprising 130 fish species classes with 28,328 image samples. The results show that our model outperforms the state-of-the-art method and significantly reduces the annotation cost. Furthermore, we benchmark our model’s performance against a public dataset (PASCAL VOC07), showcasing its effectiveness in comparison to existing methods.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信