Antara Howal, Atharva Golapkar, Yunus Khan, Siddhantha Bokade, S. Varma, Madhura Vyawahare
{"title":"基于深度卷积神经网络的手语手指拼写识别系统","authors":"Antara Howal, Atharva Golapkar, Yunus Khan, Siddhantha Bokade, S. Varma, Madhura Vyawahare","doi":"10.1109/ICNTE56631.2023.10146675","DOIUrl":null,"url":null,"abstract":"Sign Language is used by deaf and hard hearing people as it is the efficient way to communicate among themselves as well as with other people. The demand for an automated system, which converts sign language into a normal message and vice versa, has been increased to provide a normal life to people with hearing difficulties. Each sign language has a distinct character with variations in the shape of the hand, the profile of movement, and the position of the hand, face, and body parts that contribute to each character. Pattern recognition and gesture recognition are majorly used technologies for resolving the problem. In this work, real-time American Sign Language is carried out for finger-spelling recognition using Deep Convolutional Neural Network for the classification of data to bridge the gap of communication between hearing disabled people and normal people. A new dataset of more than 50 thousand images is prepared for training purposes. Outputs are based on inputs given by the user, while the designed interface provides the end user with alphabets A to Z and Numerals from 0 to 9 prediction. The overall performance based on the proposed approach is being evaluated by the use of a dataset of real-depth images captured from users. Accuracy of the system is 99.70%.","PeriodicalId":158124,"journal":{"name":"2023 5th Biennial International Conference on Nascent Technologies in Engineering (ICNTE)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Sign Language Finger-Spelling Recognition System Using Deep Convolutional Neural Network\",\"authors\":\"Antara Howal, Atharva Golapkar, Yunus Khan, Siddhantha Bokade, S. Varma, Madhura Vyawahare\",\"doi\":\"10.1109/ICNTE56631.2023.10146675\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Sign Language is used by deaf and hard hearing people as it is the efficient way to communicate among themselves as well as with other people. The demand for an automated system, which converts sign language into a normal message and vice versa, has been increased to provide a normal life to people with hearing difficulties. Each sign language has a distinct character with variations in the shape of the hand, the profile of movement, and the position of the hand, face, and body parts that contribute to each character. Pattern recognition and gesture recognition are majorly used technologies for resolving the problem. In this work, real-time American Sign Language is carried out for finger-spelling recognition using Deep Convolutional Neural Network for the classification of data to bridge the gap of communication between hearing disabled people and normal people. A new dataset of more than 50 thousand images is prepared for training purposes. Outputs are based on inputs given by the user, while the designed interface provides the end user with alphabets A to Z and Numerals from 0 to 9 prediction. The overall performance based on the proposed approach is being evaluated by the use of a dataset of real-depth images captured from users. Accuracy of the system is 99.70%.\",\"PeriodicalId\":158124,\"journal\":{\"name\":\"2023 5th Biennial International Conference on Nascent Technologies in Engineering (ICNTE)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 5th Biennial International Conference on Nascent Technologies in Engineering (ICNTE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICNTE56631.2023.10146675\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 5th Biennial International Conference on Nascent Technologies in Engineering (ICNTE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNTE56631.2023.10146675","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Sign Language Finger-Spelling Recognition System Using Deep Convolutional Neural Network
Sign Language is used by deaf and hard hearing people as it is the efficient way to communicate among themselves as well as with other people. The demand for an automated system, which converts sign language into a normal message and vice versa, has been increased to provide a normal life to people with hearing difficulties. Each sign language has a distinct character with variations in the shape of the hand, the profile of movement, and the position of the hand, face, and body parts that contribute to each character. Pattern recognition and gesture recognition are majorly used technologies for resolving the problem. In this work, real-time American Sign Language is carried out for finger-spelling recognition using Deep Convolutional Neural Network for the classification of data to bridge the gap of communication between hearing disabled people and normal people. A new dataset of more than 50 thousand images is prepared for training purposes. Outputs are based on inputs given by the user, while the designed interface provides the end user with alphabets A to Z and Numerals from 0 to 9 prediction. The overall performance based on the proposed approach is being evaluated by the use of a dataset of real-depth images captured from users. Accuracy of the system is 99.70%.