Pix2Pix-Based Grayscale Image Coloring Method

Q3 Computer Science
Hong Li, Qiaoxue Zheng, Jing Zhang, Zhuo-Ming Du, Zhanli Li, Baosheng Kang
{"title":"Pix2Pix-Based Grayscale Image Coloring Method","authors":"Hong Li, Qiaoxue Zheng, Jing Zhang, Zhuo-Ming Du, Zhanli Li, Baosheng Kang","doi":"10.3724/sp.j.1089.2021.18596","DOIUrl":null,"url":null,"abstract":": In this study, a grayscale image coloring method combining the Pix2Pix model is proposed to solve the problem of unclear object boundaries and low image coloring quality in colorization neural net-works. First, an improved U-Net structure, using eight down-sampling and up-sampling layers, is adopted to extract features and predict the image color, which improves the network model’s ability to extract deep image features. Second, the coloring image quality is tested under different loss functions, 1 L loss and smooth 1 L loss, to measure the distance between the generated image and ground truth. Finally, gradient penalty is added to improve the network stability of the training process. The gradient of each input data is penalized by constructing a new data distribution between the generated and real image distribution to limit the dis-criminator gradient. In the same experimental environment, the Pix2Pix model and summer2winter data are utilized for comparative analysis. The experiments demonstrate that the improved U-Net using the smooth 1 L loss as generator loss generates better colored images, whereas the 1 L loss better maintains the structural information of the image. Furthermore, the gradient penalty accelerates the model convergence speed, and improves the model stability and image quality. The proposed image coloring method learns deep image features and reduces the image blurs. The model raises the image quality while effectively maintaining the image structure similarity.","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"计算机辅助设计与图形学学报","FirstCategoryId":"1093","ListUrlMain":"https://doi.org/10.3724/sp.j.1089.2021.18596","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Computer Science","Score":null,"Total":0}
引用次数: 6

Abstract

: In this study, a grayscale image coloring method combining the Pix2Pix model is proposed to solve the problem of unclear object boundaries and low image coloring quality in colorization neural net-works. First, an improved U-Net structure, using eight down-sampling and up-sampling layers, is adopted to extract features and predict the image color, which improves the network model’s ability to extract deep image features. Second, the coloring image quality is tested under different loss functions, 1 L loss and smooth 1 L loss, to measure the distance between the generated image and ground truth. Finally, gradient penalty is added to improve the network stability of the training process. The gradient of each input data is penalized by constructing a new data distribution between the generated and real image distribution to limit the dis-criminator gradient. In the same experimental environment, the Pix2Pix model and summer2winter data are utilized for comparative analysis. The experiments demonstrate that the improved U-Net using the smooth 1 L loss as generator loss generates better colored images, whereas the 1 L loss better maintains the structural information of the image. Furthermore, the gradient penalty accelerates the model convergence speed, and improves the model stability and image quality. The proposed image coloring method learns deep image features and reduces the image blurs. The model raises the image quality while effectively maintaining the image structure similarity.
基于Pix2Pix的灰度图像着色方法
本研究提出了一种结合Pix2Pix模型的灰度图像着色方法,以解决着色神经网络中物体边界不清晰和图像着色质量低的问题。首先,采用改进的U-Net结构,采用8个下采样层和8个上采样层进行特征提取和图像颜色预测,提高了网络模型提取深度图像特征的能力;其次,在不同的损失函数,1 L损失和平滑1 L损失下,测试着色图像的质量,测量生成的图像与地面真值之间的距离。最后,加入梯度惩罚,提高训练过程的网络稳定性。通过在生成的图像和真实图像之间构造一个新的数据分布来限制鉴别器梯度,从而对每个输入数据的梯度进行惩罚。在相同的实验环境下,采用Pix2Pix模型和summer2winter数据进行对比分析。实验表明,采用平滑的1 L损耗作为生成损耗的改进U-Net能生成更好的彩色图像,而1 L损耗能更好地保持图像的结构信息。此外,梯度惩罚加快了模型的收敛速度,提高了模型的稳定性和图像质量。该方法学习了图像的深度特征,降低了图像的模糊程度。该模型在有效保持图像结构相似性的同时,提高了图像质量。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
计算机辅助设计与图形学学报
计算机辅助设计与图形学学报 Computer Science-Computer Graphics and Computer-Aided Design
CiteScore
1.20
自引率
0.00%
发文量
6833
期刊介绍:
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信