Est3D2Real-estimated 3D-to-real数据嵌入用于实时手语识别器

IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Kishore P.V.V. , Anil Kumar D.
{"title":"Est3D2Real-estimated 3D-to-real数据嵌入用于实时手语识别器","authors":"Kishore P.V.V. ,&nbsp;Anil Kumar D.","doi":"10.1016/j.patrec.2025.05.012","DOIUrl":null,"url":null,"abstract":"<div><div>Human pose estimation predicts 3D skeletal joints from 2D video data. These estimated 3D joints are sensitive to video data anomalies, posing a threat to applications such as real-time sign language recognition. The challenge lies in the failure of the estimation model to output pose vectors during the signing process, which significantly impacts downstream classification tasks. To address this issue, we propose the development of a lightweight estimated 3D-to-real data embedding network (Est3D2Real). This network is designed to learn the relationships between the outputs of the pose estimation framework and a 3D motion capture system. Est3D2Real is a four-layer fully connected network, consisting of one input layer, two hidden layers, and one output layer. It employs the Mean Squared Error (MSE) loss function to minimize the distance between the two modalities. The trained Est3D2Real model ensures minimal joint loss in real-time downstream classification tasks. Validation is performed on a 100-gloss 3D sign language dataset, captured using both motion capture and MediaPipe frameworks. Subsequent downstream sign classifiers built on top of the trained Est3D2Real model have shown an approximate improvement of 28%. The code with small datasets is made available at <span><span>https://github.com/pvvkishore/Est3D2Real_SL_MediaPipe_2_Motion_Capture</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"196 ","pages":"Pages 86-92"},"PeriodicalIF":3.3000,"publicationDate":"2025-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Est3D2Real-estimated 3D-to-real data embeddings for real time sign language recognizer\",\"authors\":\"Kishore P.V.V. ,&nbsp;Anil Kumar D.\",\"doi\":\"10.1016/j.patrec.2025.05.012\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Human pose estimation predicts 3D skeletal joints from 2D video data. These estimated 3D joints are sensitive to video data anomalies, posing a threat to applications such as real-time sign language recognition. The challenge lies in the failure of the estimation model to output pose vectors during the signing process, which significantly impacts downstream classification tasks. To address this issue, we propose the development of a lightweight estimated 3D-to-real data embedding network (Est3D2Real). This network is designed to learn the relationships between the outputs of the pose estimation framework and a 3D motion capture system. Est3D2Real is a four-layer fully connected network, consisting of one input layer, two hidden layers, and one output layer. It employs the Mean Squared Error (MSE) loss function to minimize the distance between the two modalities. The trained Est3D2Real model ensures minimal joint loss in real-time downstream classification tasks. Validation is performed on a 100-gloss 3D sign language dataset, captured using both motion capture and MediaPipe frameworks. Subsequent downstream sign classifiers built on top of the trained Est3D2Real model have shown an approximate improvement of 28%. The code with small datasets is made available at <span><span>https://github.com/pvvkishore/Est3D2Real_SL_MediaPipe_2_Motion_Capture</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":54638,\"journal\":{\"name\":\"Pattern Recognition Letters\",\"volume\":\"196 \",\"pages\":\"Pages 86-92\"},\"PeriodicalIF\":3.3000,\"publicationDate\":\"2025-06-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Pattern Recognition Letters\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167865525002077\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition Letters","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167865525002077","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

人体姿态估计从2D视频数据预测3D骨骼关节。这些估计的3D关节对视频数据异常很敏感,对实时手语识别等应用构成威胁。挑战在于估计模型在签名过程中无法输出姿态向量,这将严重影响下游的分类任务。为了解决这个问题,我们提出开发一个轻量级的估计3d到真实的数据嵌入网络(Est3D2Real)。该网络旨在学习姿态估计框架和3D运动捕捉系统输出之间的关系。Est3D2Real是一个四层全连接网络,由一个输入层、两个隐藏层和一个输出层组成。它采用均方误差(MSE)损失函数来最小化两个模态之间的距离。经过训练的Est3D2Real模型确保了实时下游分类任务中最小的联合损失。验证在100 gloss 3D手语数据集上执行,使用动作捕捉和MediaPipe框架捕获。随后建立在训练好的Est3D2Real模型之上的下游符号分类器显示出大约28%的改进。带有小数据集的代码可在https://github.com/pvvkishore/Est3D2Real_SL_MediaPipe_2_Motion_Capture上获得。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Est3D2Real-estimated 3D-to-real data embeddings for real time sign language recognizer
Human pose estimation predicts 3D skeletal joints from 2D video data. These estimated 3D joints are sensitive to video data anomalies, posing a threat to applications such as real-time sign language recognition. The challenge lies in the failure of the estimation model to output pose vectors during the signing process, which significantly impacts downstream classification tasks. To address this issue, we propose the development of a lightweight estimated 3D-to-real data embedding network (Est3D2Real). This network is designed to learn the relationships between the outputs of the pose estimation framework and a 3D motion capture system. Est3D2Real is a four-layer fully connected network, consisting of one input layer, two hidden layers, and one output layer. It employs the Mean Squared Error (MSE) loss function to minimize the distance between the two modalities. The trained Est3D2Real model ensures minimal joint loss in real-time downstream classification tasks. Validation is performed on a 100-gloss 3D sign language dataset, captured using both motion capture and MediaPipe frameworks. Subsequent downstream sign classifiers built on top of the trained Est3D2Real model have shown an approximate improvement of 28%. The code with small datasets is made available at https://github.com/pvvkishore/Est3D2Real_SL_MediaPipe_2_Motion_Capture.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Pattern Recognition Letters
Pattern Recognition Letters 工程技术-计算机:人工智能
CiteScore
12.40
自引率
5.90%
发文量
287
审稿时长
9.1 months
期刊介绍: Pattern Recognition Letters aims at rapid publication of concise articles of a broad interest in pattern recognition. Subject areas include all the current fields of interest represented by the Technical Committees of the International Association of Pattern Recognition, and other developing themes involving learning and recognition.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信