Generative Adversarial Networks for Visible to Infrared Video Conversion

M. S. Uddin, Jiang Li
{"title":"Generative Adversarial Networks for Visible to Infrared Video Conversion","authors":"M. S. Uddin, Jiang Li","doi":"10.5772/intechopen.93866","DOIUrl":null,"url":null,"abstract":"Deep learning models are data driven. For example, the most popular convolutional neural network (CNN) model used for image classification or object detection requires large labeled databases for training to achieve competitive performances. This requirement is not difficult to be satisfied in the visible domain since there are lots of labeled video and image databases available nowadays. However, given the less popularity of infrared (IR) camera, the availability of labeled infrared videos or image databases is limited. Therefore, training deep learning models in infrared domain is still challenging. In this chapter, we applied the pix2pix generative adversarial network (Pix2Pix GAN) and cycle-consistent GAN (Cycle GAN) models to convert visible videos to infrared videos. The Pix2Pix GAN model requires visible-infrared image pairs for training while the Cycle GAN relaxes this constraint and requires only unpaired images from both domains. We applied the two models to an open-source database where visible and infrared videos provided by the signal multimedia and telecommunications laboratory at the Federal University of Rio de Janeiro. We evaluated conversion results by performance metrics including Inception Score (IS), Frechet Inception Distance (FID) and Kernel Inception Distance (KID). Our experiments suggest that cycle-consistent GAN is more effective than pix2pix GAN for generating IR images from optical images.","PeriodicalId":171152,"journal":{"name":"Recent Advances in Image Restoration with Applications to Real World Problems","volume":"68 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Recent Advances in Image Restoration with Applications to Real World Problems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5772/intechopen.93866","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5

Abstract

Deep learning models are data driven. For example, the most popular convolutional neural network (CNN) model used for image classification or object detection requires large labeled databases for training to achieve competitive performances. This requirement is not difficult to be satisfied in the visible domain since there are lots of labeled video and image databases available nowadays. However, given the less popularity of infrared (IR) camera, the availability of labeled infrared videos or image databases is limited. Therefore, training deep learning models in infrared domain is still challenging. In this chapter, we applied the pix2pix generative adversarial network (Pix2Pix GAN) and cycle-consistent GAN (Cycle GAN) models to convert visible videos to infrared videos. The Pix2Pix GAN model requires visible-infrared image pairs for training while the Cycle GAN relaxes this constraint and requires only unpaired images from both domains. We applied the two models to an open-source database where visible and infrared videos provided by the signal multimedia and telecommunications laboratory at the Federal University of Rio de Janeiro. We evaluated conversion results by performance metrics including Inception Score (IS), Frechet Inception Distance (FID) and Kernel Inception Distance (KID). Our experiments suggest that cycle-consistent GAN is more effective than pix2pix GAN for generating IR images from optical images.
可见光到红外视频转换的生成对抗网络
深度学习模型是数据驱动的。例如,用于图像分类或目标检测的最流行的卷积神经网络(CNN)模型需要大型标记数据库进行训练以获得具有竞争力的性能。这一要求不难在可见领域得到满足,因为目前已有大量的标记视频和图像数据库。然而,由于红外摄像机的普及程度较低,标记红外视频或图像数据库的可用性是有限的。因此,在红外域训练深度学习模型仍然具有挑战性。在本章中,我们应用pix2pix生成对抗网络(pix2pix GAN)和周期一致GAN (Cycle GAN)模型将可见视频转换为红外视频。Pix2Pix GAN模型需要可见光-红外图像对进行训练,而Cycle GAN放宽了这一限制,只需要来自两个域的未配对图像。我们将这两种模型应用到一个开源数据库中,其中包含由里约热内卢联邦大学的信号多媒体和电信实验室提供的可见光和红外视频。我们通过包括Inception Score (IS)、Frechet Inception Distance (FID)和Kernel Inception Distance (KID)在内的性能指标来评估转换结果。我们的实验表明,在从光学图像生成红外图像方面,周期一致的GAN比pix2pix GAN更有效。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信