{"title":"Replay Attention and Data Augmentation Network for 3-D Face and Object Reconstruction","authors":"Zhiyuan Zhou;Lei Li;Suping Wu;Xinyu Li;Kehua Ma;Xitie Zhang","doi":"10.1109/TBIOM.2023.3261272","DOIUrl":null,"url":null,"abstract":"3D face reconstruction from single-view images plays an important role in the field of biometrics, which is a long-standing challenging problem in the wild. Traditional 3DMM-based methods directly regressed parameters, which probably caused that the network learned the discriminative informative features insufficiently. In this paper, we propose a replay attention and data augmentation network (RADAN) for 3D dense alignment and face reconstruction. Unlike the traditional attention mechanism, our replay attention module aims to increase the sensitivity of the network to informative features by adaptively recalibrating the weight response in the attention, which typically reinforces the distinguishability of the learned feature representation. In this way, the network can further improve the accuracy of face reconstruction and dense alignment in unconstrained environments. Moreover, to improve the generalization performance of the model and the ability of the network to capture local details, we present a data augmentation strategy to preprocess the sample data, which generates the images that contain more local details and occluded face in cropping and pasting manner. Furthermore, we also apply the replay attention to 3D object reconstruction task to verify the commonality of this mechanism. Extensive experimental results on widely-evaluated datasets demonstrate that our approach achieves competitive performance compared to state-of-the-art methods. Code is available at \n<uri>https://github.com/zhouzhiyuan1/RADANet</uri>\n.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"5 3","pages":"308-320"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on biometrics, behavior, and identity science","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10081017/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
3D face reconstruction from single-view images plays an important role in the field of biometrics, which is a long-standing challenging problem in the wild. Traditional 3DMM-based methods directly regressed parameters, which probably caused that the network learned the discriminative informative features insufficiently. In this paper, we propose a replay attention and data augmentation network (RADAN) for 3D dense alignment and face reconstruction. Unlike the traditional attention mechanism, our replay attention module aims to increase the sensitivity of the network to informative features by adaptively recalibrating the weight response in the attention, which typically reinforces the distinguishability of the learned feature representation. In this way, the network can further improve the accuracy of face reconstruction and dense alignment in unconstrained environments. Moreover, to improve the generalization performance of the model and the ability of the network to capture local details, we present a data augmentation strategy to preprocess the sample data, which generates the images that contain more local details and occluded face in cropping and pasting manner. Furthermore, we also apply the replay attention to 3D object reconstruction task to verify the commonality of this mechanism. Extensive experimental results on widely-evaluated datasets demonstrate that our approach achieves competitive performance compared to state-of-the-art methods. Code is available at
https://github.com/zhouzhiyuan1/RADANet
.