基于外观的唇形跟踪与克隆说话面部

Bouchra Abboud, G. Chollet
{"title":"基于外观的唇形跟踪与克隆说话面部","authors":"Bouchra Abboud, G. Chollet","doi":"10.1109/ISPA.2005.195427","DOIUrl":null,"url":null,"abstract":"This paper addresses the issue of lip tracking and cloning using appearance models. In this perspective, a statistical color-based lip detector is first used to classify the pixels of an image into lip and non-lip pixels. This rough estimation of the lip position is then used to initialize an appearance model. This models convergence allows to refine the positions of the MPEG-4 compatible feature points placed around the lip contours. The optimal position of the feature points is then used as a first estimate to compute the position at the next frame of an image sequence to perform speaking lip tracking. To animate an unknown face image in such a way that it reproduces the lip motion of the driving sequence a piecewise affine transform is applied forcing the target lips feature points to match the automatically detected feature points of each frame of the training sequence to perform lip motion cloning. Preliminary results show that the synthetic talking faces obtained are photorealistic and reproduce accurately the movements of the tracked mouth.","PeriodicalId":238993,"journal":{"name":"ISPA 2005. Proceedings of the 4th International Symposium on Image and Signal Processing and Analysis, 2005.","volume":"41 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2005-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"12","resultStr":"{\"title\":\"Appearance based lip tracking and cloning on speaking faces\",\"authors\":\"Bouchra Abboud, G. Chollet\",\"doi\":\"10.1109/ISPA.2005.195427\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper addresses the issue of lip tracking and cloning using appearance models. In this perspective, a statistical color-based lip detector is first used to classify the pixels of an image into lip and non-lip pixels. This rough estimation of the lip position is then used to initialize an appearance model. This models convergence allows to refine the positions of the MPEG-4 compatible feature points placed around the lip contours. The optimal position of the feature points is then used as a first estimate to compute the position at the next frame of an image sequence to perform speaking lip tracking. To animate an unknown face image in such a way that it reproduces the lip motion of the driving sequence a piecewise affine transform is applied forcing the target lips feature points to match the automatically detected feature points of each frame of the training sequence to perform lip motion cloning. Preliminary results show that the synthetic talking faces obtained are photorealistic and reproduce accurately the movements of the tracked mouth.\",\"PeriodicalId\":238993,\"journal\":{\"name\":\"ISPA 2005. Proceedings of the 4th International Symposium on Image and Signal Processing and Analysis, 2005.\",\"volume\":\"41 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2005-10-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"12\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ISPA 2005. Proceedings of the 4th International Symposium on Image and Signal Processing and Analysis, 2005.\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISPA.2005.195427\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPA 2005. Proceedings of the 4th International Symposium on Image and Signal Processing and Analysis, 2005.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISPA.2005.195427","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 12

摘要

本文讨论了使用外观模型进行唇形跟踪和克隆的问题。从这个角度来看,首先使用基于统计颜色的唇形检测器将图像像素分类为唇形和非唇形像素。然后使用唇位置的粗略估计来初始化外观模型。该模型的收敛性允许细化放置在唇轮廓周围的MPEG-4兼容特征点的位置。然后使用特征点的最佳位置作为第一个估计来计算图像序列下一帧的位置,以执行说话唇跟踪。为了使未知人脸图像以再现驱动序列的嘴唇运动的方式动画化,应用分段仿射变换迫使目标嘴唇特征点与训练序列的每帧自动检测的特征点匹配,以执行嘴唇运动克隆。初步结果表明,合成的说话脸具有较好的真实感,能较准确地再现被跟踪嘴部的运动。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Appearance based lip tracking and cloning on speaking faces
This paper addresses the issue of lip tracking and cloning using appearance models. In this perspective, a statistical color-based lip detector is first used to classify the pixels of an image into lip and non-lip pixels. This rough estimation of the lip position is then used to initialize an appearance model. This models convergence allows to refine the positions of the MPEG-4 compatible feature points placed around the lip contours. The optimal position of the feature points is then used as a first estimate to compute the position at the next frame of an image sequence to perform speaking lip tracking. To animate an unknown face image in such a way that it reproduces the lip motion of the driving sequence a piecewise affine transform is applied forcing the target lips feature points to match the automatically detected feature points of each frame of the training sequence to perform lip motion cloning. Preliminary results show that the synthetic talking faces obtained are photorealistic and reproduce accurately the movements of the tracked mouth.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信