{"title":"Stacked Generative Adversarial Networks for Image Generation based on U-Net discriminator","authors":"Wanyan Feng, Zuqiang Meng, L. Wang","doi":"10.1109/CACML55074.2022.00132","DOIUrl":null,"url":null,"abstract":"Although Generative Adversarial Networks (GANs) are powerful generative models and have shown remarkable success in various tasks recently but suffers from generating high-quality images. In this paper, we proposed a U-Net-based discriminator structure on the network structure of the two-stage Stacked Generative Adversarial Networks (StackGAN++), aiming to generate high-resolution images with actual shapes and textures. To gain more insight from limited datasets, we focused on improving the discriminator's ability to discriminate the real from the fake. The discriminator based on U-Net architecture allows providing details per-pixels and global feedback to the generator to maintain the global coherence of synthetic images and the realistic of local shape and textures. In addition, for the problem that the training effect on a small number of sample datasets is not ide-al, we further improve the quality of the generated samples by transfer learning of model parameters. Compared with the StackGAN++ baseline, experiments show that we have significantly improved the IS and FID evaluation indicators of the ImageNet subset.","PeriodicalId":137505,"journal":{"name":"2022 Asia Conference on Algorithms, Computing and Machine Learning (CACML)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 Asia Conference on Algorithms, Computing and Machine Learning (CACML)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CACML55074.2022.00132","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Although Generative Adversarial Networks (GANs) are powerful generative models and have shown remarkable success in various tasks recently but suffers from generating high-quality images. In this paper, we proposed a U-Net-based discriminator structure on the network structure of the two-stage Stacked Generative Adversarial Networks (StackGAN++), aiming to generate high-resolution images with actual shapes and textures. To gain more insight from limited datasets, we focused on improving the discriminator's ability to discriminate the real from the fake. The discriminator based on U-Net architecture allows providing details per-pixels and global feedback to the generator to maintain the global coherence of synthetic images and the realistic of local shape and textures. In addition, for the problem that the training effect on a small number of sample datasets is not ide-al, we further improve the quality of the generated samples by transfer learning of model parameters. Compared with the StackGAN++ baseline, experiments show that we have significantly improved the IS and FID evaluation indicators of the ImageNet subset.