{"title":"Semi-Supervised Semantic Segmentation of Class-Imbalanced Images: A Hierarchical Self-Attention Generative Adversarial Network","authors":"Lu Chai, Qinyuan Liu","doi":"10.1109/ICIVC55077.2022.9886496","DOIUrl":null,"url":null,"abstract":"How to train models with unlabeled data and implement one trained model across several data sets are key problems in computer vision applications that require high-cost annotations. Recently, a generative model [1] proves its advantages in semi-supervised segmentation and out-of-domain generalization. However, this method becomes less effective when meet with class-imbalanced images whose foreground occupies small areas. To solve this problem, we introduce a hierarchical generative model with a self-attention mechanism to help with capturing features of foreground objects. Concretely, we apply a two-stage hierarchical generative model to perform image synthesis with the self-attention mechanism. Since attention maps are also semantic labels in segmentation fields, the hierarchical self-attention model can synthesize images and corresponding segmentation labels simultaneously. At test time, the segmentation is achieved by mapping input images into latent presentations with two encoders and synthesizing labels with the generative model. We evaluate our hierarchical model on three biomedical segmentation data sets. The experimental results demonstrate that our method outperforms other baselines on semi-supervised segmentation of class-imbalanced images, and meanwhile, pre-serves out-of-domain generalization ability.","PeriodicalId":227073,"journal":{"name":"2022 7th International Conference on Image, Vision and Computing (ICIVC)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 7th International Conference on Image, Vision and Computing (ICIVC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICIVC55077.2022.9886496","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
How to train models with unlabeled data and implement one trained model across several data sets are key problems in computer vision applications that require high-cost annotations. Recently, a generative model [1] proves its advantages in semi-supervised segmentation and out-of-domain generalization. However, this method becomes less effective when meet with class-imbalanced images whose foreground occupies small areas. To solve this problem, we introduce a hierarchical generative model with a self-attention mechanism to help with capturing features of foreground objects. Concretely, we apply a two-stage hierarchical generative model to perform image synthesis with the self-attention mechanism. Since attention maps are also semantic labels in segmentation fields, the hierarchical self-attention model can synthesize images and corresponding segmentation labels simultaneously. At test time, the segmentation is achieved by mapping input images into latent presentations with two encoders and synthesizing labels with the generative model. We evaluate our hierarchical model on three biomedical segmentation data sets. The experimental results demonstrate that our method outperforms other baselines on semi-supervised segmentation of class-imbalanced images, and meanwhile, pre-serves out-of-domain generalization ability.