A U-Net Architecture for Inpainting Lightstage Normal Maps

Comput. Pub Date : 2024-02-19 DOI:10.3390/computers13020056
Hancheng Zuo, Bernard Tiddeman
{"title":"A U-Net Architecture for Inpainting Lightstage Normal Maps","authors":"Hancheng Zuo, Bernard Tiddeman","doi":"10.3390/computers13020056","DOIUrl":null,"url":null,"abstract":"In this paper, we investigate the inpainting of normal maps that were captured from a lightstage. Occlusion of parts of the face during performance capture can be caused by the movement of, e.g., arms, hair, or props. Inpainting is the process of interpolating missing areas of an image with plausible data. We build on previous works about general image inpainting that use generative adversarial networks (GANs). We extend our previous work on normal map inpainting to use a U-Net structured generator network. Our method takes into account the nature of the normal map data and so requires modification of the loss function. We use a cosine loss rather than the more common mean squared error loss when training the generator. Due to the small amount of training data available, even when using synthetic datasets, we require significant augmentation, which also needs to take account of the particular nature of the input data. Image flipping and inplane rotations need to properly flip and rotate the normal vectors. During training, we monitor key performance metrics including the average loss, structural similarity index measure (SSIM), and peak signal-to-noise ratio (PSNR) of the generator, alongside the average loss and accuracy of the discriminator. Our analysis reveals that the proposed model generates high-quality, realistic inpainted normal maps, demonstrating the potential for application to performance capture. The results of this investigation provide a baseline on which future researchers can build with more advanced networks and comparison with inpainting of the source images used to generate the normal maps.","PeriodicalId":10526,"journal":{"name":"Comput.","volume":"186 ","pages":"56"},"PeriodicalIF":0.0000,"publicationDate":"2024-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Comput.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3390/computers13020056","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In this paper, we investigate the inpainting of normal maps that were captured from a lightstage. Occlusion of parts of the face during performance capture can be caused by the movement of, e.g., arms, hair, or props. Inpainting is the process of interpolating missing areas of an image with plausible data. We build on previous works about general image inpainting that use generative adversarial networks (GANs). We extend our previous work on normal map inpainting to use a U-Net structured generator network. Our method takes into account the nature of the normal map data and so requires modification of the loss function. We use a cosine loss rather than the more common mean squared error loss when training the generator. Due to the small amount of training data available, even when using synthetic datasets, we require significant augmentation, which also needs to take account of the particular nature of the input data. Image flipping and inplane rotations need to properly flip and rotate the normal vectors. During training, we monitor key performance metrics including the average loss, structural similarity index measure (SSIM), and peak signal-to-noise ratio (PSNR) of the generator, alongside the average loss and accuracy of the discriminator. Our analysis reveals that the proposed model generates high-quality, realistic inpainted normal maps, demonstrating the potential for application to performance capture. The results of this investigation provide a baseline on which future researchers can build with more advanced networks and comparison with inpainting of the source images used to generate the normal maps.
用于绘制光场法线映射的 U-Net 架构
在本文中,我们研究了如何对从灯光舞台捕捉到的法线贴图进行内绘。在捕捉表演过程中,手臂、头发或道具等的移动可能会导致面部部分遮挡。内绘是用可信数据对图像缺失区域进行插值的过程。我们在以往使用生成式对抗网络(GAN)进行一般图像内绘的基础上进行了改进。我们扩展了之前的法线贴图绘制工作,使用了 U-Net 结构生成器网络。我们的方法考虑到了法线贴图数据的性质,因此需要修改损失函数。在训练生成器时,我们使用余弦损失而不是更常见的均方误差损失。由于可用的训练数据较少,即使使用合成数据集,我们也需要大量的增强数据,这也需要考虑输入数据的特殊性质。图像翻转和平面内旋转需要正确翻转和旋转法向量。在训练过程中,我们对关键性能指标进行监控,包括生成器的平均损失、结构相似性指数(SSIM)和峰值信噪比(PSNR),以及判别器的平均损失和准确性。我们的分析表明,所提出的模型能生成高质量、逼真的内绘法线图,显示了应用于性能捕捉的潜力。这项研究的结果提供了一个基线,未来的研究人员可以在此基础上建立更先进的网络,并对用于生成法线图的源图像进行染色比较。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信