Caminate Na Rang , Paulo Jerónimo , Carlos Mora , Sandra Jardim
{"title":"Hand Gesture Recognition using Machine Learning","authors":"Caminate Na Rang , Paulo Jerónimo , Carlos Mora , Sandra Jardim","doi":"10.1016/j.procs.2025.02.112","DOIUrl":null,"url":null,"abstract":"<div><div>Sign language recognition is a growing area of research, with applications ranging from gestural communication to controlling devices using gestures. One of the challenges inherent to sign language recognition is the ability to translate gestures into meaningful information, such as letters, words or even sentences. Machine Learning, which has emerged as a powerful tool for solving a wide variety of complex problems, namely in the field of computer vision, plays a key role, enabling computers to understand and interpret complex gestures. In this paper, we present a Machine Learning model focused on classifying hand gestures that represent the letters of the Latin alphabet. The objective of this work is to create a solution capable of accurately identifying which letter of the Latin alphabet is being represented by a hand gesture in an image. To classify manual gestures was used the Random Forest Machine Learning classification model, which is fed with the vector of features extracted from the region of interest in the image. To implement the proposed approach, a database of RGB images of hand gestures was created. To extract the characteristics of the gestures, was used the MediaPipe open source framework. The solution presents hand gesture classification precisions by class ranging between 98.8% and 74.4%, with an accuracy of 92.3%, that represents an improvement over previous approaches.</div></div>","PeriodicalId":20465,"journal":{"name":"Procedia Computer Science","volume":"256 ","pages":"Pages 198-205"},"PeriodicalIF":0.0000,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Procedia Computer Science","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1877050925004697","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Sign language recognition is a growing area of research, with applications ranging from gestural communication to controlling devices using gestures. One of the challenges inherent to sign language recognition is the ability to translate gestures into meaningful information, such as letters, words or even sentences. Machine Learning, which has emerged as a powerful tool for solving a wide variety of complex problems, namely in the field of computer vision, plays a key role, enabling computers to understand and interpret complex gestures. In this paper, we present a Machine Learning model focused on classifying hand gestures that represent the letters of the Latin alphabet. The objective of this work is to create a solution capable of accurately identifying which letter of the Latin alphabet is being represented by a hand gesture in an image. To classify manual gestures was used the Random Forest Machine Learning classification model, which is fed with the vector of features extracted from the region of interest in the image. To implement the proposed approach, a database of RGB images of hand gestures was created. To extract the characteristics of the gestures, was used the MediaPipe open source framework. The solution presents hand gesture classification precisions by class ranging between 98.8% and 74.4%, with an accuracy of 92.3%, that represents an improvement over previous approaches.