Hong Li, Qiaoxue Zheng, Jing Zhang, Zhuo-Ming Du, Zhanli Li, Baosheng Kang
{"title":"Pix2Pix-Based Grayscale Image Coloring Method","authors":"Hong Li, Qiaoxue Zheng, Jing Zhang, Zhuo-Ming Du, Zhanli Li, Baosheng Kang","doi":"10.3724/sp.j.1089.2021.18596","DOIUrl":null,"url":null,"abstract":": In this study, a grayscale image coloring method combining the Pix2Pix model is proposed to solve the problem of unclear object boundaries and low image coloring quality in colorization neural net-works. First, an improved U-Net structure, using eight down-sampling and up-sampling layers, is adopted to extract features and predict the image color, which improves the network model’s ability to extract deep image features. Second, the coloring image quality is tested under different loss functions, 1 L loss and smooth 1 L loss, to measure the distance between the generated image and ground truth. Finally, gradient penalty is added to improve the network stability of the training process. The gradient of each input data is penalized by constructing a new data distribution between the generated and real image distribution to limit the dis-criminator gradient. In the same experimental environment, the Pix2Pix model and summer2winter data are utilized for comparative analysis. The experiments demonstrate that the improved U-Net using the smooth 1 L loss as generator loss generates better colored images, whereas the 1 L loss better maintains the structural information of the image. Furthermore, the gradient penalty accelerates the model convergence speed, and improves the model stability and image quality. The proposed image coloring method learns deep image features and reduces the image blurs. The model raises the image quality while effectively maintaining the image structure similarity.","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"计算机辅助设计与图形学学报","FirstCategoryId":"1093","ListUrlMain":"https://doi.org/10.3724/sp.j.1089.2021.18596","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Computer Science","Score":null,"Total":0}
引用次数: 6
Abstract
: In this study, a grayscale image coloring method combining the Pix2Pix model is proposed to solve the problem of unclear object boundaries and low image coloring quality in colorization neural net-works. First, an improved U-Net structure, using eight down-sampling and up-sampling layers, is adopted to extract features and predict the image color, which improves the network model’s ability to extract deep image features. Second, the coloring image quality is tested under different loss functions, 1 L loss and smooth 1 L loss, to measure the distance between the generated image and ground truth. Finally, gradient penalty is added to improve the network stability of the training process. The gradient of each input data is penalized by constructing a new data distribution between the generated and real image distribution to limit the dis-criminator gradient. In the same experimental environment, the Pix2Pix model and summer2winter data are utilized for comparative analysis. The experiments demonstrate that the improved U-Net using the smooth 1 L loss as generator loss generates better colored images, whereas the 1 L loss better maintains the structural information of the image. Furthermore, the gradient penalty accelerates the model convergence speed, and improves the model stability and image quality. The proposed image coloring method learns deep image features and reduces the image blurs. The model raises the image quality while effectively maintaining the image structure similarity.