Hong Li, Qiaoxue Zheng, Jing Zhang, Zhuo-Ming Du, Zhanli Li, Baosheng Kang
{"title":"基于Pix2Pix的灰度图像着色方法","authors":"Hong Li, Qiaoxue Zheng, Jing Zhang, Zhuo-Ming Du, Zhanli Li, Baosheng Kang","doi":"10.3724/sp.j.1089.2021.18596","DOIUrl":null,"url":null,"abstract":": In this study, a grayscale image coloring method combining the Pix2Pix model is proposed to solve the problem of unclear object boundaries and low image coloring quality in colorization neural net-works. First, an improved U-Net structure, using eight down-sampling and up-sampling layers, is adopted to extract features and predict the image color, which improves the network model’s ability to extract deep image features. Second, the coloring image quality is tested under different loss functions, 1 L loss and smooth 1 L loss, to measure the distance between the generated image and ground truth. Finally, gradient penalty is added to improve the network stability of the training process. The gradient of each input data is penalized by constructing a new data distribution between the generated and real image distribution to limit the dis-criminator gradient. In the same experimental environment, the Pix2Pix model and summer2winter data are utilized for comparative analysis. The experiments demonstrate that the improved U-Net using the smooth 1 L loss as generator loss generates better colored images, whereas the 1 L loss better maintains the structural information of the image. Furthermore, the gradient penalty accelerates the model convergence speed, and improves the model stability and image quality. The proposed image coloring method learns deep image features and reduces the image blurs. The model raises the image quality while effectively maintaining the image structure similarity.","PeriodicalId":52442,"journal":{"name":"计算机辅助设计与图形学学报","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"Pix2Pix-Based Grayscale Image Coloring Method\",\"authors\":\"Hong Li, Qiaoxue Zheng, Jing Zhang, Zhuo-Ming Du, Zhanli Li, Baosheng Kang\",\"doi\":\"10.3724/sp.j.1089.2021.18596\",\"DOIUrl\":null,\"url\":null,\"abstract\":\": In this study, a grayscale image coloring method combining the Pix2Pix model is proposed to solve the problem of unclear object boundaries and low image coloring quality in colorization neural net-works. First, an improved U-Net structure, using eight down-sampling and up-sampling layers, is adopted to extract features and predict the image color, which improves the network model’s ability to extract deep image features. Second, the coloring image quality is tested under different loss functions, 1 L loss and smooth 1 L loss, to measure the distance between the generated image and ground truth. Finally, gradient penalty is added to improve the network stability of the training process. The gradient of each input data is penalized by constructing a new data distribution between the generated and real image distribution to limit the dis-criminator gradient. In the same experimental environment, the Pix2Pix model and summer2winter data are utilized for comparative analysis. The experiments demonstrate that the improved U-Net using the smooth 1 L loss as generator loss generates better colored images, whereas the 1 L loss better maintains the structural information of the image. Furthermore, the gradient penalty accelerates the model convergence speed, and improves the model stability and image quality. The proposed image coloring method learns deep image features and reduces the image blurs. The model raises the image quality while effectively maintaining the image structure similarity.\",\"PeriodicalId\":52442,\"journal\":{\"name\":\"计算机辅助设计与图形学学报\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"计算机辅助设计与图形学学报\",\"FirstCategoryId\":\"1093\",\"ListUrlMain\":\"https://doi.org/10.3724/sp.j.1089.2021.18596\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"Computer Science\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"计算机辅助设计与图形学学报","FirstCategoryId":"1093","ListUrlMain":"https://doi.org/10.3724/sp.j.1089.2021.18596","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Computer Science","Score":null,"Total":0}
: In this study, a grayscale image coloring method combining the Pix2Pix model is proposed to solve the problem of unclear object boundaries and low image coloring quality in colorization neural net-works. First, an improved U-Net structure, using eight down-sampling and up-sampling layers, is adopted to extract features and predict the image color, which improves the network model’s ability to extract deep image features. Second, the coloring image quality is tested under different loss functions, 1 L loss and smooth 1 L loss, to measure the distance between the generated image and ground truth. Finally, gradient penalty is added to improve the network stability of the training process. The gradient of each input data is penalized by constructing a new data distribution between the generated and real image distribution to limit the dis-criminator gradient. In the same experimental environment, the Pix2Pix model and summer2winter data are utilized for comparative analysis. The experiments demonstrate that the improved U-Net using the smooth 1 L loss as generator loss generates better colored images, whereas the 1 L loss better maintains the structural information of the image. Furthermore, the gradient penalty accelerates the model convergence speed, and improves the model stability and image quality. The proposed image coloring method learns deep image features and reduces the image blurs. The model raises the image quality while effectively maintaining the image structure similarity.