Kan Li, Cuiwei Liu, Chong Du, Zhuo Yan, Zhaokui Li, Xiangbin Shi
{"title":"Learning from coarsely-labeled images for semantic segmentation","authors":"Kan Li, Cuiwei Liu, Chong Du, Zhuo Yan, Zhaokui Li, Xiangbin Shi","doi":"10.1109/YAC57282.2022.10023665","DOIUrl":null,"url":null,"abstract":"Image segmentation aims to assign a semantic label to each pixel of an image and thus requires accurate pixel-wise annotations of massive images to train high-performance models. This paper aims to alleviate the arduous annotations by learning from coarser polygonal annotations that can be acquired at a lower cost. A novel self-training framework is proposed to build a robust semantic segmentation model in the presence of noisy labels. A superpixel-based relabeling strategy is developed to refine the original annotations according to the local context. Then multiple segmentation models are trained on the revised annotations and produce multiple pseudo labels for model re-training. Finally, a new segmentation model is re-trained by fusing multiple pseudo-labels in terms of their confidences. Experimental results on the Cityscapes dataset demonstrate the effectiveness of the proposed method.","PeriodicalId":272227,"journal":{"name":"2022 37th Youth Academic Annual Conference of Chinese Association of Automation (YAC)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 37th Youth Academic Annual Conference of Chinese Association of Automation (YAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/YAC57282.2022.10023665","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Image segmentation aims to assign a semantic label to each pixel of an image and thus requires accurate pixel-wise annotations of massive images to train high-performance models. This paper aims to alleviate the arduous annotations by learning from coarser polygonal annotations that can be acquired at a lower cost. A novel self-training framework is proposed to build a robust semantic segmentation model in the presence of noisy labels. A superpixel-based relabeling strategy is developed to refine the original annotations according to the local context. Then multiple segmentation models are trained on the revised annotations and produce multiple pseudo labels for model re-training. Finally, a new segmentation model is re-trained by fusing multiple pseudo-labels in terms of their confidences. Experimental results on the Cityscapes dataset demonstrate the effectiveness of the proposed method.