{"title":"Information-Cycle Dual Contractive Learning with Regularization for Unsupervised Image-to-Image Translation","authors":"Xin Sun, Zhuang Yin, Siqi Zhang","doi":"10.1109/ISAIEE57420.2022.00047","DOIUrl":null,"url":null,"abstract":"In the tasks of image-to-image translation, we aim to add the appearance features of the target domain on the basis of retaining the structural features of the input image. Contrastive learning methods are gradually applied to image translation tasks. Despite the previous progress in image translation models, it remains challenging to employ an efficient learning setting by capturing image features from multiple perspectives to build mappings. In this paper, we propose a novel method based on information-cycle dual contrastive learning with regularization (ICDCLGAN-R) to improve the overall quality and visual observability which mainly consists of three components: dual-domain GAN, contrastive regularization (CR) and information cycle. Specifically, dual-domain GAN contributes to learning cross-domain image mappings and capturing the domain gap more efficiently by maximizing mutual information between corresponding patches of input and output. Simultaneously, CR utilizes two images from one domain (one real image and one generated image) and one generated image belonging to another domain as positive, anchor and negative, respectively, aiming to pull positive pairs in some metric space and push apart the representation between negative pairs. CR compares multi-source image features meanwhile enhancing information reusability. Besides, we put generated images into dual-domain GAN ulteriorly to perform additional iteration. It reinforces and verifies reconfigurability between images from different domains. In our experiments, we demonstrate that our method improves the quality of generated images and utilizes image features more efficiently.","PeriodicalId":345703,"journal":{"name":"2022 International Symposium on Advances in Informatics, Electronics and Education (ISAIEE)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Symposium on Advances in Informatics, Electronics and Education (ISAIEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISAIEE57420.2022.00047","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In the tasks of image-to-image translation, we aim to add the appearance features of the target domain on the basis of retaining the structural features of the input image. Contrastive learning methods are gradually applied to image translation tasks. Despite the previous progress in image translation models, it remains challenging to employ an efficient learning setting by capturing image features from multiple perspectives to build mappings. In this paper, we propose a novel method based on information-cycle dual contrastive learning with regularization (ICDCLGAN-R) to improve the overall quality and visual observability which mainly consists of three components: dual-domain GAN, contrastive regularization (CR) and information cycle. Specifically, dual-domain GAN contributes to learning cross-domain image mappings and capturing the domain gap more efficiently by maximizing mutual information between corresponding patches of input and output. Simultaneously, CR utilizes two images from one domain (one real image and one generated image) and one generated image belonging to another domain as positive, anchor and negative, respectively, aiming to pull positive pairs in some metric space and push apart the representation between negative pairs. CR compares multi-source image features meanwhile enhancing information reusability. Besides, we put generated images into dual-domain GAN ulteriorly to perform additional iteration. It reinforces and verifies reconfigurability between images from different domains. In our experiments, we demonstrate that our method improves the quality of generated images and utilizes image features more efficiently.