{"title":"Faster convergence and reduction of overfitting in numerical hand sign recognition using DCNN","authors":"A. Tushar, Akm Ashiquzzaman, Md. Rashedul Islam","doi":"10.1109/R10-HTC.2017.8289040","DOIUrl":null,"url":null,"abstract":"Hand signs and signals are the staple form of expression for the hearing and speech impaired people. Human Computer Interaction technology enable people to interact with computer machine using hand gestures. Common sign languages use separate hand signals to communicate different decimals. Recent developments in Deep Convolutional Neural Networks (DCNN) have opened the door to recognize and classify this visual form of gestures more accurately. In this paper, a layer-wise optimized neural network architecture is proposed where batch normalization contributes to faster convergence of training, and introduction of dropout technique mitigates data overfitting. Batch normalization forces each training batch toward zero mean and unit variance, leading to improved flow of gradients through the model and convergence in shorter time. Dropout forces neurons of neural network to regularize, resulting in reduced overfitting. A constructed numerical hand gesture data set is used for validating the claims based on American Sign Language system. The proposed model is shown to surpass other methods in classifying these numerical hand signs successfully.","PeriodicalId":411099,"journal":{"name":"2017 IEEE Region 10 Humanitarian Technology Conference (R10-HTC)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"12","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 IEEE Region 10 Humanitarian Technology Conference (R10-HTC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/R10-HTC.2017.8289040","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 12
Abstract
Hand signs and signals are the staple form of expression for the hearing and speech impaired people. Human Computer Interaction technology enable people to interact with computer machine using hand gestures. Common sign languages use separate hand signals to communicate different decimals. Recent developments in Deep Convolutional Neural Networks (DCNN) have opened the door to recognize and classify this visual form of gestures more accurately. In this paper, a layer-wise optimized neural network architecture is proposed where batch normalization contributes to faster convergence of training, and introduction of dropout technique mitigates data overfitting. Batch normalization forces each training batch toward zero mean and unit variance, leading to improved flow of gradients through the model and convergence in shorter time. Dropout forces neurons of neural network to regularize, resulting in reduced overfitting. A constructed numerical hand gesture data set is used for validating the claims based on American Sign Language system. The proposed model is shown to surpass other methods in classifying these numerical hand signs successfully.