{"title":"基于外观的唇形跟踪与克隆说话面部","authors":"Bouchra Abboud, G. Chollet","doi":"10.1109/ISPA.2005.195427","DOIUrl":null,"url":null,"abstract":"This paper addresses the issue of lip tracking and cloning using appearance models. In this perspective, a statistical color-based lip detector is first used to classify the pixels of an image into lip and non-lip pixels. This rough estimation of the lip position is then used to initialize an appearance model. This models convergence allows to refine the positions of the MPEG-4 compatible feature points placed around the lip contours. The optimal position of the feature points is then used as a first estimate to compute the position at the next frame of an image sequence to perform speaking lip tracking. To animate an unknown face image in such a way that it reproduces the lip motion of the driving sequence a piecewise affine transform is applied forcing the target lips feature points to match the automatically detected feature points of each frame of the training sequence to perform lip motion cloning. Preliminary results show that the synthetic talking faces obtained are photorealistic and reproduce accurately the movements of the tracked mouth.","PeriodicalId":238993,"journal":{"name":"ISPA 2005. Proceedings of the 4th International Symposium on Image and Signal Processing and Analysis, 2005.","volume":"41 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2005-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"12","resultStr":"{\"title\":\"Appearance based lip tracking and cloning on speaking faces\",\"authors\":\"Bouchra Abboud, G. Chollet\",\"doi\":\"10.1109/ISPA.2005.195427\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper addresses the issue of lip tracking and cloning using appearance models. In this perspective, a statistical color-based lip detector is first used to classify the pixels of an image into lip and non-lip pixels. This rough estimation of the lip position is then used to initialize an appearance model. This models convergence allows to refine the positions of the MPEG-4 compatible feature points placed around the lip contours. The optimal position of the feature points is then used as a first estimate to compute the position at the next frame of an image sequence to perform speaking lip tracking. To animate an unknown face image in such a way that it reproduces the lip motion of the driving sequence a piecewise affine transform is applied forcing the target lips feature points to match the automatically detected feature points of each frame of the training sequence to perform lip motion cloning. Preliminary results show that the synthetic talking faces obtained are photorealistic and reproduce accurately the movements of the tracked mouth.\",\"PeriodicalId\":238993,\"journal\":{\"name\":\"ISPA 2005. Proceedings of the 4th International Symposium on Image and Signal Processing and Analysis, 2005.\",\"volume\":\"41 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2005-10-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"12\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ISPA 2005. Proceedings of the 4th International Symposium on Image and Signal Processing and Analysis, 2005.\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISPA.2005.195427\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPA 2005. Proceedings of the 4th International Symposium on Image and Signal Processing and Analysis, 2005.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISPA.2005.195427","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Appearance based lip tracking and cloning on speaking faces
This paper addresses the issue of lip tracking and cloning using appearance models. In this perspective, a statistical color-based lip detector is first used to classify the pixels of an image into lip and non-lip pixels. This rough estimation of the lip position is then used to initialize an appearance model. This models convergence allows to refine the positions of the MPEG-4 compatible feature points placed around the lip contours. The optimal position of the feature points is then used as a first estimate to compute the position at the next frame of an image sequence to perform speaking lip tracking. To animate an unknown face image in such a way that it reproduces the lip motion of the driving sequence a piecewise affine transform is applied forcing the target lips feature points to match the automatically detected feature points of each frame of the training sequence to perform lip motion cloning. Preliminary results show that the synthetic talking faces obtained are photorealistic and reproduce accurately the movements of the tracked mouth.