{"title":"Recognition of Real-Time Hand Gestures using Mediapipe Holistic Model and LSTM with MLP Architecture","authors":"Maricel L. Amit, Arnel C. Fajardo, Ruji P. Medina","doi":"10.1109/ICSPC55597.2022.10001800","DOIUrl":null,"url":null,"abstract":"This study used computer vision to capture real-time hand gestures using the MediaPipe Holistic Model and LSTM with MLP architecture to bridge the communication gap between the hearing majority and the deaf minority. The structure of LSTM architecture was consist of 84 neurons with a 30, 1662 input vector that will be transmitted to the MLP model. It has five layers with corresponding nodes of 84, 56, 28,14, and 7 and a dropout with rate values of 0.4 in the first, second, and third layers, while 0.5 in fourth layer. The proposed method was trained and validated with 1000 epoch which achieved a 100 percent accuracy rate in the recognition of real-time hand gesture. Additionally, the researchers envisioned to incorporate hand gesture detection into a number of real-world applications for human-computer interaction in the coming years.","PeriodicalId":334831,"journal":{"name":"2022 IEEE 10th Conference on Systems, Process & Control (ICSPC)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 10th Conference on Systems, Process & Control (ICSPC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSPC55597.2022.10001800","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
This study used computer vision to capture real-time hand gestures using the MediaPipe Holistic Model and LSTM with MLP architecture to bridge the communication gap between the hearing majority and the deaf minority. The structure of LSTM architecture was consist of 84 neurons with a 30, 1662 input vector that will be transmitted to the MLP model. It has five layers with corresponding nodes of 84, 56, 28,14, and 7 and a dropout with rate values of 0.4 in the first, second, and third layers, while 0.5 in fourth layer. The proposed method was trained and validated with 1000 epoch which achieved a 100 percent accuracy rate in the recognition of real-time hand gesture. Additionally, the researchers envisioned to incorporate hand gesture detection into a number of real-world applications for human-computer interaction in the coming years.