Abdullah Azeem, Waqar Riaz, Abubakar Siddique, Tahir Junaid Saifullah
{"title":"使用生成对抗网络的人类视频合成","authors":"Abdullah Azeem, Waqar Riaz, Abubakar Siddique, Tahir Junaid Saifullah","doi":"10.1117/12.2574615","DOIUrl":null,"url":null,"abstract":"In this work, a video synthesis model based on Generative Adversarial Networks (Human GAN) is proposed, whose objective is to generate a photorealistic output by learning the mapping function from an input source to output video. However, the image to image generation is a quite popular problem, but the video synthesis problem is still unexplored. Directly employing existing image generation method without taking temporal dynamics into account leads to frequent temporally incoherent output with low visual quality. The proposed approach solves this problem by wisely designing generators and discriminators combined with Spatio-temporal adversarial objects. While comparing it to some robust baselines on public benchmarks, the proposed model proves to be superior in generating temporally coherent videos with extremely low artifacts. And results achieved by the proposed model are more realistic on both quantitative and qualitative measures compared to other existing baselines techniques.","PeriodicalId":90079,"journal":{"name":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","volume":"415 14","pages":"115260I - 115260I-5"},"PeriodicalIF":0.0000,"publicationDate":"2020-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Human video synthesis using generative adversarial networks\",\"authors\":\"Abdullah Azeem, Waqar Riaz, Abubakar Siddique, Tahir Junaid Saifullah\",\"doi\":\"10.1117/12.2574615\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this work, a video synthesis model based on Generative Adversarial Networks (Human GAN) is proposed, whose objective is to generate a photorealistic output by learning the mapping function from an input source to output video. However, the image to image generation is a quite popular problem, but the video synthesis problem is still unexplored. Directly employing existing image generation method without taking temporal dynamics into account leads to frequent temporally incoherent output with low visual quality. The proposed approach solves this problem by wisely designing generators and discriminators combined with Spatio-temporal adversarial objects. While comparing it to some robust baselines on public benchmarks, the proposed model proves to be superior in generating temporally coherent videos with extremely low artifacts. And results achieved by the proposed model are more realistic on both quantitative and qualitative measures compared to other existing baselines techniques.\",\"PeriodicalId\":90079,\"journal\":{\"name\":\"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging\",\"volume\":\"415 14\",\"pages\":\"115260I - 115260I-5\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-06-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1117/12.2574615\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"... International Workshop on Pattern Recognition in NeuroImaging. International Workshop on Pattern Recognition in NeuroImaging","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.2574615","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Human video synthesis using generative adversarial networks
In this work, a video synthesis model based on Generative Adversarial Networks (Human GAN) is proposed, whose objective is to generate a photorealistic output by learning the mapping function from an input source to output video. However, the image to image generation is a quite popular problem, but the video synthesis problem is still unexplored. Directly employing existing image generation method without taking temporal dynamics into account leads to frequent temporally incoherent output with low visual quality. The proposed approach solves this problem by wisely designing generators and discriminators combined with Spatio-temporal adversarial objects. While comparing it to some robust baselines on public benchmarks, the proposed model proves to be superior in generating temporally coherent videos with extremely low artifacts. And results achieved by the proposed model are more realistic on both quantitative and qualitative measures compared to other existing baselines techniques.