基于时间正则化正则分量分析的手势语视频手势语识别方法

Shohei Tanaka, A. Okazaki, N. Kato, H. Hino, K. Fukui
{"title":"基于时间正则化正则分量分析的手势语视频手势语识别方法","authors":"Shohei Tanaka, A. Okazaki, N. Kato, H. Hino, K. Fukui","doi":"10.1109/ISBA.2016.7477238","DOIUrl":null,"url":null,"abstract":"A method for spotting specific words in sign language video is proposed. In classes and talks given using Japanese Sign Language, words that do not have a defined sign, such as the names of people, objects, and places, are represented by sets of multiple characters from the Japanese finger alphabet. The difficulty of recognizing these words has created strong demand for the ability to spot specific words in order to help interpreters and the audience to follow a talk. We address the spotting task by employing the basic idea of temporal regularized canonical correlation analysis (TRCCA), which can simultaneously handle shape and motion information about a 3D object. The classification accuracy of TRCCA is enhanced by incorporating two functions: 1) parallel processing with multiple time scales, 2) strong implicit feature mapping by nonlinear orthogonalization. The enhanced TRCCA is called \"kernel orthogonal TRCCA (KOTRCCA)\". The effectiveness of the proposed method using KOTRCCA is demonstrated through experiments spotting eight different words in sign language videos.","PeriodicalId":198009,"journal":{"name":"2016 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Spotting fingerspelled words from sign language video by temporally regularized canonical component analysis\",\"authors\":\"Shohei Tanaka, A. Okazaki, N. Kato, H. Hino, K. Fukui\",\"doi\":\"10.1109/ISBA.2016.7477238\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A method for spotting specific words in sign language video is proposed. In classes and talks given using Japanese Sign Language, words that do not have a defined sign, such as the names of people, objects, and places, are represented by sets of multiple characters from the Japanese finger alphabet. The difficulty of recognizing these words has created strong demand for the ability to spot specific words in order to help interpreters and the audience to follow a talk. We address the spotting task by employing the basic idea of temporal regularized canonical correlation analysis (TRCCA), which can simultaneously handle shape and motion information about a 3D object. The classification accuracy of TRCCA is enhanced by incorporating two functions: 1) parallel processing with multiple time scales, 2) strong implicit feature mapping by nonlinear orthogonalization. The enhanced TRCCA is called \\\"kernel orthogonal TRCCA (KOTRCCA)\\\". The effectiveness of the proposed method using KOTRCCA is demonstrated through experiments spotting eight different words in sign language videos.\",\"PeriodicalId\":198009,\"journal\":{\"name\":\"2016 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA)\",\"volume\":\"25 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-02-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISBA.2016.7477238\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE International Conference on Identity, Security and Behavior Analysis (ISBA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISBA.2016.7477238","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

摘要

提出了一种识别手语视频中特定单词的方法。在使用日语手语进行的课堂和讲座中,没有明确符号的单词,如人名、物体和地点的名称,由日语手指字母表中的多个字符组合来表示。识别这些单词的困难产生了对识别特定单词的能力的强烈需求,以帮助口译员和听众跟上谈话。我们采用时间正则化典型相关分析(TRCCA)的基本思想来解决识别任务,该方法可以同时处理三维物体的形状和运动信息。利用多时间尺度并行处理和非线性正交化的强隐式特征映射两种功能,提高了TRCCA的分类精度。增强的TRCCA被称为“核正交TRCCA (KOTRCCA)”。通过在手语视频中识别8个不同单词的实验,验证了该方法的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Spotting fingerspelled words from sign language video by temporally regularized canonical component analysis
A method for spotting specific words in sign language video is proposed. In classes and talks given using Japanese Sign Language, words that do not have a defined sign, such as the names of people, objects, and places, are represented by sets of multiple characters from the Japanese finger alphabet. The difficulty of recognizing these words has created strong demand for the ability to spot specific words in order to help interpreters and the audience to follow a talk. We address the spotting task by employing the basic idea of temporal regularized canonical correlation analysis (TRCCA), which can simultaneously handle shape and motion information about a 3D object. The classification accuracy of TRCCA is enhanced by incorporating two functions: 1) parallel processing with multiple time scales, 2) strong implicit feature mapping by nonlinear orthogonalization. The enhanced TRCCA is called "kernel orthogonal TRCCA (KOTRCCA)". The effectiveness of the proposed method using KOTRCCA is demonstrated through experiments spotting eight different words in sign language videos.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信