{"title":"Automatic Hand Gesture Recognition with Semantic Segmentation and Deep Learning","authors":"Bristy Chanda, H. Nyeem","doi":"10.1109/icaeee54957.2022.9836425","DOIUrl":null,"url":null,"abstract":"Automatic Hand Gesture Recognition is a key requirement for variety of applications, including translation of Sign Language, Human-Computer Interaction (HCI) and, ubiquitous vision-based systems. Due to the lighting variance and complicated background in the input image set of gestures, meeting this criterion remains a challenge. This paper introduces semantic segmentation to deep learning-based hand gesture recognition system for sign language translation. Building on the U - Net architecture, the proposed model obtains the semantically segmented mask of the input image, which is then fed to convolutional neural networks (CNNs) for multiclass classification. The proposed model is trained and tested for four different depths of the CNN architectures followed by the comparison with some pre-trained CNN architectures such as Inception V3, VGG16, VGG19, ResNet50. The proposed model is evaluated on National University of Singapore (NUS) hand posture dataset II (subset A), which contains 2000 images in 10 classes. A significant recognition rate of 97.15 % is achieved for the proposed model outperforming a set of prominent models and demonstrating its promises for sign language translation.","PeriodicalId":383872,"journal":{"name":"2022 International Conference on Advancement in Electrical and Electronic Engineering (ICAEEE)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-02-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Advancement in Electrical and Electronic Engineering (ICAEEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/icaeee54957.2022.9836425","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Automatic Hand Gesture Recognition is a key requirement for variety of applications, including translation of Sign Language, Human-Computer Interaction (HCI) and, ubiquitous vision-based systems. Due to the lighting variance and complicated background in the input image set of gestures, meeting this criterion remains a challenge. This paper introduces semantic segmentation to deep learning-based hand gesture recognition system for sign language translation. Building on the U - Net architecture, the proposed model obtains the semantically segmented mask of the input image, which is then fed to convolutional neural networks (CNNs) for multiclass classification. The proposed model is trained and tested for four different depths of the CNN architectures followed by the comparison with some pre-trained CNN architectures such as Inception V3, VGG16, VGG19, ResNet50. The proposed model is evaluated on National University of Singapore (NUS) hand posture dataset II (subset A), which contains 2000 images in 10 classes. A significant recognition rate of 97.15 % is achieved for the proposed model outperforming a set of prominent models and demonstrating its promises for sign language translation.