A Deep Learning Framework to Reconstruct Face under Mask

Gourango Modak, S. Das, Md. Ajharul Islam Miraj, Md. Kishor Morol
{"title":"A Deep Learning Framework to Reconstruct Face under Mask","authors":"Gourango Modak, S. Das, Md. Ajharul Islam Miraj, Md. Kishor Morol","doi":"10.1109/CDMA54072.2022.00038","DOIUrl":null,"url":null,"abstract":"While deep learning-based image reconstruction methods have shown significant success in removing objects from pictures, they have yet to achieve acceptable results for attributing consistency to gender, ethnicity, expression, and other characteristics like the topological structure of the face. The purpose of this work is to extract the mask region from a masked image and rebuild the area that has been detected. This problem is complex because (i) it is difficult to determine the gender of an image hidden behind a mask, which causes the network to become confused and reconstruct the male face as a female or vice versa; (ii) we may receive images from multiple angles, making it extremely difficult to maintain the actual shape, topological structure of the face and a natural image; and (iii) there are problems with various mask forms because, in some cases, the area of the mask cannot be anticipated precisely; certain parts of the mask remain on the face after completion. To solve this complex task, we split the problem into three phases: landmark detection, object detection for the targeted mask area, and inpainting the addressed mask region. To begin, to solve the first problem, we have used gender classification, which detects the actual gender behind a mask, then we detect the landmark of the masked facial image. Second, we identified the non-face item, i.e., the mask, and used the Mask R-CNN network to create the binary mask of the observed mask area. Thirdly, we developed an inpainting network that uses anticipated landmarks to create realistic images. To segment the mask, this article uses a mask R-CNN and offers a binary segmentation map for identifying the mask area. Additionally, we generated the image utilizing landmarks as structural guidance through a GAN-based network. The studies presented in this paper use the FFHQ and CelebA datasets. This study outperformed all prior studies in terms of generating cutting-edge results for real-world pictures gathered from the web.","PeriodicalId":313042,"journal":{"name":"2022 7th International Conference on Data Science and Machine Learning Applications (CDMA)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 7th International Conference on Data Science and Machine Learning Applications (CDMA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CDMA54072.2022.00038","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8

Abstract

While deep learning-based image reconstruction methods have shown significant success in removing objects from pictures, they have yet to achieve acceptable results for attributing consistency to gender, ethnicity, expression, and other characteristics like the topological structure of the face. The purpose of this work is to extract the mask region from a masked image and rebuild the area that has been detected. This problem is complex because (i) it is difficult to determine the gender of an image hidden behind a mask, which causes the network to become confused and reconstruct the male face as a female or vice versa; (ii) we may receive images from multiple angles, making it extremely difficult to maintain the actual shape, topological structure of the face and a natural image; and (iii) there are problems with various mask forms because, in some cases, the area of the mask cannot be anticipated precisely; certain parts of the mask remain on the face after completion. To solve this complex task, we split the problem into three phases: landmark detection, object detection for the targeted mask area, and inpainting the addressed mask region. To begin, to solve the first problem, we have used gender classification, which detects the actual gender behind a mask, then we detect the landmark of the masked facial image. Second, we identified the non-face item, i.e., the mask, and used the Mask R-CNN network to create the binary mask of the observed mask area. Thirdly, we developed an inpainting network that uses anticipated landmarks to create realistic images. To segment the mask, this article uses a mask R-CNN and offers a binary segmentation map for identifying the mask area. Additionally, we generated the image utilizing landmarks as structural guidance through a GAN-based network. The studies presented in this paper use the FFHQ and CelebA datasets. This study outperformed all prior studies in terms of generating cutting-edge results for real-world pictures gathered from the web.
面具下人脸重构的深度学习框架
虽然基于深度学习的图像重建方法在从图像中删除物体方面取得了重大成功,但在将一致性归因于性别、种族、表情和其他特征(如面部拓扑结构)方面,它们尚未取得可接受的结果。本工作的目的是从被遮挡的图像中提取掩模区域,并重建已检测到的区域。这个问题很复杂,因为(i)很难确定隐藏在面具后面的图像的性别,这导致网络变得困惑,并将男性面部重建为女性面部,反之亦然;(ii)我们可能从多个角度接收图像,因此很难保持面部的实际形状、拓扑结构和自然图像;(iii)各种口罩形式存在问题,因为在某些情况下,口罩的面积无法精确预测;完成后,口罩的某些部分仍留在脸上。为了解决这个复杂的任务,我们将问题分为三个阶段:地标检测,目标遮罩区域的目标检测,以及对寻址遮罩区域进行涂漆。首先,为了解决第一个问题,我们使用了性别分类,它检测面具后面的实际性别,然后我们检测被面具的面部图像的地标。其次,我们识别非人脸项目,即掩码,并使用mask R-CNN网络创建观察到的掩码区域的二进制掩码。第三,我们开发了一个绘图网络,使用预期的地标来创建逼真的图像。为了分割掩码,本文使用掩码R-CNN,并提供用于识别掩码区域的二值分割图。此外,我们通过基于gan的网络利用地标作为结构指导生成图像。本文的研究使用了FFHQ和CelebA数据集。这项研究在生成从网络上收集的真实世界图片的尖端结果方面优于所有先前的研究。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信