Xiangcheng Du, Zhao Zhou, Yanlong Wang, Zhuoyao Wang, Yingbin Zheng, Cheng Jin
{"title":"MultiColor: Image Colorization by Learning from Multiple Color Spaces","authors":"Xiangcheng Du, Zhao Zhou, Yanlong Wang, Zhuoyao Wang, Yingbin Zheng, Cheng Jin","doi":"arxiv-2408.04172","DOIUrl":null,"url":null,"abstract":"Deep networks have shown impressive performance in the image restoration\ntasks, such as image colorization. However, we find that previous approaches\nrely on the digital representation from single color model with a specific\nmapping function, a.k.a., color space, during the colorization pipeline. In\nthis paper, we first investigate the modeling of different color spaces, and\nfind each of them exhibiting distinctive characteristics with unique\ndistribution of colors. The complementarity among multiple color spaces leads\nto benefits for the image colorization task. We present MultiColor, a new learning-based approach to automatically\ncolorize grayscale images that combines clues from multiple color spaces.\nSpecifically, we employ a set of dedicated colorization modules for individual\ncolor space. Within each module, a transformer decoder is first employed to\nrefine color query embeddings and then a color mapper produces color channel\nprediction using the embeddings and semantic features. With these predicted\ncolor channels representing various color spaces, a complementary network is\ndesigned to exploit the complementarity and generate pleasing and reasonable\ncolorized images. We conduct extensive experiments on real-world datasets, and\nthe results demonstrate superior performance over the state-of-the-arts.","PeriodicalId":501480,"journal":{"name":"arXiv - CS - Multimedia","volume":"11 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Multimedia","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.04172","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Deep networks have shown impressive performance in the image restoration
tasks, such as image colorization. However, we find that previous approaches
rely on the digital representation from single color model with a specific
mapping function, a.k.a., color space, during the colorization pipeline. In
this paper, we first investigate the modeling of different color spaces, and
find each of them exhibiting distinctive characteristics with unique
distribution of colors. The complementarity among multiple color spaces leads
to benefits for the image colorization task. We present MultiColor, a new learning-based approach to automatically
colorize grayscale images that combines clues from multiple color spaces.
Specifically, we employ a set of dedicated colorization modules for individual
color space. Within each module, a transformer decoder is first employed to
refine color query embeddings and then a color mapper produces color channel
prediction using the embeddings and semantic features. With these predicted
color channels representing various color spaces, a complementary network is
designed to exploit the complementarity and generate pleasing and reasonable
colorized images. We conduct extensive experiments on real-world datasets, and
the results demonstrate superior performance over the state-of-the-arts.