{"title":"在潜在空间中嵌入多种特征进行人脸属性编辑","authors":"Rui Yuan, Xiping He, Dan He, Yue Li","doi":"10.1117/12.2667748","DOIUrl":null,"url":null,"abstract":"Face attribute editing, one of the important research directions in face image synthesis and processing techniques, aims to photorealistic editing single or multiple attributes of face images on demand using editing and generation models. Most existing methods are based on generative adversarial networks, using target attribute vectors to control the editing region or Gaussian noise as conditional input to capture texture details. However, these cannot better control the consistency of attributes in irrelevant regions, while the generation of fidelity is also limited. In this paper, we propose a method that uses an optimized latent space to fuse the attribute feature maps into the latent space. At the same time, make full use of the conditional information for additional constraints. Then, in the image generation phase, we use a progressive architecture for controlled editing of face attributes at different granularities. At last, we also conducted an ablation study on the selected training scheme further to demonstrate the stability and accuracy of our chosen method. The experiments show that our proposed approach, using an end-to-end progressive image translation network architecture, obtained qualitative (FID) as well as quantitative (LPIPS) face image editing results.","PeriodicalId":345723,"journal":{"name":"Fifth International Conference on Computer Information Science and Artificial Intelligence","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Embedding diverse features in latent space for face attribute editing\",\"authors\":\"Rui Yuan, Xiping He, Dan He, Yue Li\",\"doi\":\"10.1117/12.2667748\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Face attribute editing, one of the important research directions in face image synthesis and processing techniques, aims to photorealistic editing single or multiple attributes of face images on demand using editing and generation models. Most existing methods are based on generative adversarial networks, using target attribute vectors to control the editing region or Gaussian noise as conditional input to capture texture details. However, these cannot better control the consistency of attributes in irrelevant regions, while the generation of fidelity is also limited. In this paper, we propose a method that uses an optimized latent space to fuse the attribute feature maps into the latent space. At the same time, make full use of the conditional information for additional constraints. Then, in the image generation phase, we use a progressive architecture for controlled editing of face attributes at different granularities. At last, we also conducted an ablation study on the selected training scheme further to demonstrate the stability and accuracy of our chosen method. The experiments show that our proposed approach, using an end-to-end progressive image translation network architecture, obtained qualitative (FID) as well as quantitative (LPIPS) face image editing results.\",\"PeriodicalId\":345723,\"journal\":{\"name\":\"Fifth International Conference on Computer Information Science and Artificial Intelligence\",\"volume\":\"14 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-03-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Fifth International Conference on Computer Information Science and Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1117/12.2667748\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Fifth International Conference on Computer Information Science and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.2667748","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Embedding diverse features in latent space for face attribute editing
Face attribute editing, one of the important research directions in face image synthesis and processing techniques, aims to photorealistic editing single or multiple attributes of face images on demand using editing and generation models. Most existing methods are based on generative adversarial networks, using target attribute vectors to control the editing region or Gaussian noise as conditional input to capture texture details. However, these cannot better control the consistency of attributes in irrelevant regions, while the generation of fidelity is also limited. In this paper, we propose a method that uses an optimized latent space to fuse the attribute feature maps into the latent space. At the same time, make full use of the conditional information for additional constraints. Then, in the image generation phase, we use a progressive architecture for controlled editing of face attributes at different granularities. At last, we also conducted an ablation study on the selected training scheme further to demonstrate the stability and accuracy of our chosen method. The experiments show that our proposed approach, using an end-to-end progressive image translation network architecture, obtained qualitative (FID) as well as quantitative (LPIPS) face image editing results.