Shohei Tanaka, A. Okazaki, N. Kato, H. Hino, K. Fukui
{"title":"基于时间正则化正则分量分析的手势语视频手势语识别方法","authors":"Shohei Tanaka, A. Okazaki, N. Kato, H. Hino, K. Fukui","doi":"10.1109/ISBA.2016.7477238","DOIUrl":null,"url":null,"abstract":"A method for spotting specific words in sign language video is proposed. In classes and talks given using Japanese Sign Language, words that do not have a defined sign, such as the names of people, objects, and places, are represented by sets of multiple characters from the Japanese finger alphabet. The difficulty of recognizing these words has created strong demand for the ability to spot specific words in order to help interpreters and the audience to follow a talk. We address the spotting task by employing the basic idea of temporal regularized canonical correlation analysis (TRCCA), which can simultaneously handle shape and motion information about a 3D object. The classification accuracy of TRCCA is enhanced by incorporating two functions: 1) parallel processing with multiple time scales, 2) strong implicit feature mapping by nonlinear orthogonalization. The enhanced TRCCA is called \"kernel orthogonal TRCCA (KOTRCCA)\". The effectiveness of the proposed method using KOTRCCA is demonstrated through experiments spotting eight different words in sign language videos.","PeriodicalId":198009,"journal":{"name":"2016 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Spotting fingerspelled words from sign language video by temporally regularized canonical component analysis\",\"authors\":\"Shohei Tanaka, A. Okazaki, N. Kato, H. Hino, K. Fukui\",\"doi\":\"10.1109/ISBA.2016.7477238\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A method for spotting specific words in sign language video is proposed. In classes and talks given using Japanese Sign Language, words that do not have a defined sign, such as the names of people, objects, and places, are represented by sets of multiple characters from the Japanese finger alphabet. The difficulty of recognizing these words has created strong demand for the ability to spot specific words in order to help interpreters and the audience to follow a talk. We address the spotting task by employing the basic idea of temporal regularized canonical correlation analysis (TRCCA), which can simultaneously handle shape and motion information about a 3D object. The classification accuracy of TRCCA is enhanced by incorporating two functions: 1) parallel processing with multiple time scales, 2) strong implicit feature mapping by nonlinear orthogonalization. The enhanced TRCCA is called \\\"kernel orthogonal TRCCA (KOTRCCA)\\\". The effectiveness of the proposed method using KOTRCCA is demonstrated through experiments spotting eight different words in sign language videos.\",\"PeriodicalId\":198009,\"journal\":{\"name\":\"2016 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA)\",\"volume\":\"25 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-02-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISBA.2016.7477238\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISBA.2016.7477238","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Spotting fingerspelled words from sign language video by temporally regularized canonical component analysis
A method for spotting specific words in sign language video is proposed. In classes and talks given using Japanese Sign Language, words that do not have a defined sign, such as the names of people, objects, and places, are represented by sets of multiple characters from the Japanese finger alphabet. The difficulty of recognizing these words has created strong demand for the ability to spot specific words in order to help interpreters and the audience to follow a talk. We address the spotting task by employing the basic idea of temporal regularized canonical correlation analysis (TRCCA), which can simultaneously handle shape and motion information about a 3D object. The classification accuracy of TRCCA is enhanced by incorporating two functions: 1) parallel processing with multiple time scales, 2) strong implicit feature mapping by nonlinear orthogonalization. The enhanced TRCCA is called "kernel orthogonal TRCCA (KOTRCCA)". The effectiveness of the proposed method using KOTRCCA is demonstrated through experiments spotting eight different words in sign language videos.