{"title":"3D人脸点云超分辨率网络","authors":"Jiaxin Li, Feiyu Zhu, X. Yang, Qijun Zhao","doi":"10.1109/IJCB52358.2021.9484379","DOIUrl":null,"url":null,"abstract":"With the development of consumer-level depth sensors, 3D face point cloud data can be easily captured now. However, such data are often accompanied by low resolution, noise, and holes. At the same time, high-precision 3D scanners are bulky and can not be widely used in daily applications due to costs and inconvenience. To fill the gap between low and high resolution 3D faces, we propose a two-stage framework named the face point cloud super-resolution network (FPSRN) to recover high-resolution 3D face data from the low-resolution counterparts. As the human faces can be aligned into a unified coordinate system, we formulate point cloud super-resolution as a z-coordinate prediction problem. Cascaded auto-encoders are employed to retain both global structure and boundary information of different face regions during super-resolution. Compared with state- of-the-art point cloud completion methods and depth estimation methods, our method improves the Earth-Mover’s Distance (EMD) and the Root Mean Square Error (RMSE) metrics by 43% and 25%, respectively.","PeriodicalId":175984,"journal":{"name":"2021 IEEE International Joint Conference on Biometrics (IJCB)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"3D Face Point Cloud Super-Resolution Network\",\"authors\":\"Jiaxin Li, Feiyu Zhu, X. Yang, Qijun Zhao\",\"doi\":\"10.1109/IJCB52358.2021.9484379\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the development of consumer-level depth sensors, 3D face point cloud data can be easily captured now. However, such data are often accompanied by low resolution, noise, and holes. At the same time, high-precision 3D scanners are bulky and can not be widely used in daily applications due to costs and inconvenience. To fill the gap between low and high resolution 3D faces, we propose a two-stage framework named the face point cloud super-resolution network (FPSRN) to recover high-resolution 3D face data from the low-resolution counterparts. As the human faces can be aligned into a unified coordinate system, we formulate point cloud super-resolution as a z-coordinate prediction problem. Cascaded auto-encoders are employed to retain both global structure and boundary information of different face regions during super-resolution. Compared with state- of-the-art point cloud completion methods and depth estimation methods, our method improves the Earth-Mover’s Distance (EMD) and the Root Mean Square Error (RMSE) metrics by 43% and 25%, respectively.\",\"PeriodicalId\":175984,\"journal\":{\"name\":\"2021 IEEE International Joint Conference on Biometrics (IJCB)\",\"volume\":\"25 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-08-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE International Joint Conference on Biometrics (IJCB)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IJCB52358.2021.9484379\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Joint Conference on Biometrics (IJCB)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCB52358.2021.9484379","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
With the development of consumer-level depth sensors, 3D face point cloud data can be easily captured now. However, such data are often accompanied by low resolution, noise, and holes. At the same time, high-precision 3D scanners are bulky and can not be widely used in daily applications due to costs and inconvenience. To fill the gap between low and high resolution 3D faces, we propose a two-stage framework named the face point cloud super-resolution network (FPSRN) to recover high-resolution 3D face data from the low-resolution counterparts. As the human faces can be aligned into a unified coordinate system, we formulate point cloud super-resolution as a z-coordinate prediction problem. Cascaded auto-encoders are employed to retain both global structure and boundary information of different face regions during super-resolution. Compared with state- of-the-art point cloud completion methods and depth estimation methods, our method improves the Earth-Mover’s Distance (EMD) and the Root Mean Square Error (RMSE) metrics by 43% and 25%, respectively.