Information-Cycle Dual Contractive Learning with Regularization for Unsupervised Image-to-Image Translation

Xin Sun, Zhuang Yin, Siqi Zhang
{"title":"Information-Cycle Dual Contractive Learning with Regularization for Unsupervised Image-to-Image Translation","authors":"Xin Sun, Zhuang Yin, Siqi Zhang","doi":"10.1109/ISAIEE57420.2022.00047","DOIUrl":null,"url":null,"abstract":"In the tasks of image-to-image translation, we aim to add the appearance features of the target domain on the basis of retaining the structural features of the input image. Contrastive learning methods are gradually applied to image translation tasks. Despite the previous progress in image translation models, it remains challenging to employ an efficient learning setting by capturing image features from multiple perspectives to build mappings. In this paper, we propose a novel method based on information-cycle dual contrastive learning with regularization (ICDCLGAN-R) to improve the overall quality and visual observability which mainly consists of three components: dual-domain GAN, contrastive regularization (CR) and information cycle. Specifically, dual-domain GAN contributes to learning cross-domain image mappings and capturing the domain gap more efficiently by maximizing mutual information between corresponding patches of input and output. Simultaneously, CR utilizes two images from one domain (one real image and one generated image) and one generated image belonging to another domain as positive, anchor and negative, respectively, aiming to pull positive pairs in some metric space and push apart the representation between negative pairs. CR compares multi-source image features meanwhile enhancing information reusability. Besides, we put generated images into dual-domain GAN ulteriorly to perform additional iteration. It reinforces and verifies reconfigurability between images from different domains. In our experiments, we demonstrate that our method improves the quality of generated images and utilizes image features more efficiently.","PeriodicalId":345703,"journal":{"name":"2022 International Symposium on Advances in Informatics, Electronics and Education (ISAIEE)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Symposium on Advances in Informatics, Electronics and Education (ISAIEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISAIEE57420.2022.00047","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In the tasks of image-to-image translation, we aim to add the appearance features of the target domain on the basis of retaining the structural features of the input image. Contrastive learning methods are gradually applied to image translation tasks. Despite the previous progress in image translation models, it remains challenging to employ an efficient learning setting by capturing image features from multiple perspectives to build mappings. In this paper, we propose a novel method based on information-cycle dual contrastive learning with regularization (ICDCLGAN-R) to improve the overall quality and visual observability which mainly consists of three components: dual-domain GAN, contrastive regularization (CR) and information cycle. Specifically, dual-domain GAN contributes to learning cross-domain image mappings and capturing the domain gap more efficiently by maximizing mutual information between corresponding patches of input and output. Simultaneously, CR utilizes two images from one domain (one real image and one generated image) and one generated image belonging to another domain as positive, anchor and negative, respectively, aiming to pull positive pairs in some metric space and push apart the representation between negative pairs. CR compares multi-source image features meanwhile enhancing information reusability. Besides, we put generated images into dual-domain GAN ulteriorly to perform additional iteration. It reinforces and verifies reconfigurability between images from different domains. In our experiments, we demonstrate that our method improves the quality of generated images and utilizes image features more efficiently.
无监督图像到图像翻译的正则化信息循环双收缩学习
在图像到图像的翻译任务中,我们的目标是在保留输入图像的结构特征的基础上添加目标域的外观特征。对比学习方法逐渐被应用到图像翻译任务中。尽管先前在图像翻译模型方面取得了进展,但通过从多个角度捕获图像特征来构建映射,采用有效的学习设置仍然具有挑战性。本文提出了一种基于信息循环的正则化双对比学习(ICDCLGAN-R)方法,该方法主要由双域GAN、对比正则化(CR)和信息循环三个部分组成,以提高图像的整体质量和视觉可观察性。具体而言,双域GAN通过最大化输入和输出对应patch之间的互信息,有助于学习跨域图像映射并更有效地捕获域间隙。同时,CR利用一个域的两幅图像(一幅实景图像和一幅生成图像)和另一个域的一幅生成图像分别作为正、锚和负,目的是在某个度量空间中拉动正对,拉开负对之间的表示。CR对多源图像特征进行比较,同时增强信息的可重用性。此外,我们将生成的图像进一步放入双域GAN中进行额外迭代。它加强并验证了来自不同域的图像之间的可重构性。在实验中,我们证明了我们的方法提高了生成图像的质量,并更有效地利用了图像特征。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信