Maahin Rathinagiriswaran, Swapneel Managaokar, K. R. Yashaskara Jois, Kartik Vijaykumar Suvarna, Niranjana Krupa
{"title":"Inflated 3D Architecture for South Indian Sign Language Recognition","authors":"Maahin Rathinagiriswaran, Swapneel Managaokar, K. R. Yashaskara Jois, Kartik Vijaykumar Suvarna, Niranjana Krupa","doi":"10.1109/ICMNWC52512.2021.9688520","DOIUrl":null,"url":null,"abstract":"The inability to speak is considered to be a true disability that affects both the speaker and the listener. Therefore, there is a need for a method that can recognize sign languages so that a successful communication is established. This paper proposes a method that classifies South-Indian Sign Language and translates it in real-time to Kannada language. The first few samples for the dataset were obtained from the official ISL website while the rest of the samples were manually recorded. Optical flow method is employed to extract the motion features in a video and the resulting frames are used as input to the proposed Inflated-3D model which recognizes the sign language. Stratified k-fold cross- validation is used to improve the performance. The predicted sign in text form is then translated to Kannada language through the use of GoogleTrans API and further synthesized into a speech segment using the GTTS open-source library. In addition, a comparative study of the proposed methodology with other techniques that have been proposed to recognize sign languages through video streams has been presented. The proposed method resulted in an average accuracy of 0.8709.","PeriodicalId":186283,"journal":{"name":"2021 IEEE International Conference on Mobile Networks and Wireless Communications (ICMNWC)","volume":"333 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Mobile Networks and Wireless Communications (ICMNWC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMNWC52512.2021.9688520","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The inability to speak is considered to be a true disability that affects both the speaker and the listener. Therefore, there is a need for a method that can recognize sign languages so that a successful communication is established. This paper proposes a method that classifies South-Indian Sign Language and translates it in real-time to Kannada language. The first few samples for the dataset were obtained from the official ISL website while the rest of the samples were manually recorded. Optical flow method is employed to extract the motion features in a video and the resulting frames are used as input to the proposed Inflated-3D model which recognizes the sign language. Stratified k-fold cross- validation is used to improve the performance. The predicted sign in text form is then translated to Kannada language through the use of GoogleTrans API and further synthesized into a speech segment using the GTTS open-source library. In addition, a comparative study of the proposed methodology with other techniques that have been proposed to recognize sign languages through video streams has been presented. The proposed method resulted in an average accuracy of 0.8709.