{"title":"Extraction of frame-difference features based on PCA and ICA for lip-reading","authors":"K. Lee, M. Lee, Soo-Young Lee","doi":"10.1109/IJCNN.2005.1555835","DOIUrl":null,"url":null,"abstract":"The features of human lip motion from video clips are extracted by principal component analysis (PCA) and independent component analysis (ICA). Unlike many other features extracted from single-frame static images or multi-frame dynamic images, we extracted the features from the differences of consecutive frames. The PCA results in global features, while local features are extracted by the ICA. The features are extracted from several consecutive multi-frame differences as well as single-frame differences. The dynamic nature of multi-frame differences is more eminent. The resulting features maybe applicable in lip-reading and synthesis of lip motion videos with text-to-speech capability.","PeriodicalId":365690,"journal":{"name":"Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005.","volume":"122 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2005-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"9","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN.2005.1555835","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 9
Abstract
The features of human lip motion from video clips are extracted by principal component analysis (PCA) and independent component analysis (ICA). Unlike many other features extracted from single-frame static images or multi-frame dynamic images, we extracted the features from the differences of consecutive frames. The PCA results in global features, while local features are extracted by the ICA. The features are extracted from several consecutive multi-frame differences as well as single-frame differences. The dynamic nature of multi-frame differences is more eminent. The resulting features maybe applicable in lip-reading and synthesis of lip motion videos with text-to-speech capability.