{"title":"UCR-SSL: Uncertainty-Based Consistency Regularization for Semi-Supervised Learning","authors":"Seungil Lee, Hyun Kim, Dayoung Chun","doi":"10.1109/ICEIC57457.2023.10049938","DOIUrl":null,"url":null,"abstract":"Recently, semi-supervised learning methods are being actively developed to increase the performance of neural networks by using large amounts of unlabeled data. Among these techniques, pseudo-labeling methods have the advantage of low computational complexity, but are vulnerable to missing annotations. To solve this problem, we propose a method called uncertainty-based consistency regularization (UCR). UCR models a detection head to obtain different outputs for input images and computes a feature map of each. Subsequently, these feature maps are matched with the original and filtered ground truth (GT), and are classified as positive and negative samples, respectively. In this process, missing samples are generated by the filtered GT; therefore, we use a specialized loss function designed to reduce the logit difference of the samples for robustness against missing annotations. We also use the uncertainty extracted through Gaussian modeling as a criterion for annotation filtering to train the network to focus on reliable results. As a result of experiments with an SSD model on the Pascal VOC dataset, the proposed approach achieved an improvement of 0.7% in terms of mAP compared to a baseline method.","PeriodicalId":373752,"journal":{"name":"2023 International Conference on Electronics, Information, and Communication (ICEIC)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Electronics, Information, and Communication (ICEIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICEIC57457.2023.10049938","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Recently, semi-supervised learning methods are being actively developed to increase the performance of neural networks by using large amounts of unlabeled data. Among these techniques, pseudo-labeling methods have the advantage of low computational complexity, but are vulnerable to missing annotations. To solve this problem, we propose a method called uncertainty-based consistency regularization (UCR). UCR models a detection head to obtain different outputs for input images and computes a feature map of each. Subsequently, these feature maps are matched with the original and filtered ground truth (GT), and are classified as positive and negative samples, respectively. In this process, missing samples are generated by the filtered GT; therefore, we use a specialized loss function designed to reduce the logit difference of the samples for robustness against missing annotations. We also use the uncertainty extracted through Gaussian modeling as a criterion for annotation filtering to train the network to focus on reliable results. As a result of experiments with an SSD model on the Pascal VOC dataset, the proposed approach achieved an improvement of 0.7% in terms of mAP compared to a baseline method.
近年来,人们正在积极开发半监督学习方法,通过使用大量未标记数据来提高神经网络的性能。在这些技术中,伪标注方法具有计算复杂度低的优点,但容易出现标注缺失的问题。为了解决这个问题,我们提出了一种基于不确定性的一致性正则化(UCR)方法。UCR对检测头进行建模,得到输入图像的不同输出,并计算每个输出的特征图。随后,将这些特征映射与原始和过滤后的ground truth (GT)进行匹配,并分别分类为正样本和负样本。在这个过程中,缺失样本由过滤后的GT生成;因此,我们使用专门的损失函数来减少样本的logit差异,以增强对缺失注释的鲁棒性。我们还使用通过高斯建模提取的不确定性作为注释过滤的标准,以训练网络关注可靠的结果。在Pascal VOC数据集上使用SSD模型进行的实验结果表明,与基线方法相比,所提出的方法在mAP方面实现了0.7%的改进。