{"title":"基于深度学习的语言障碍人士眼手语交流系统。","authors":"Rajesh Kannan Megalingam, Sakthiprasad Kuttankulangara Manoharan, Gokul Riju, Shree Rajesh Raagul Vadivel","doi":"10.1080/17483107.2025.2532698","DOIUrl":null,"url":null,"abstract":"<p><p><b>Objective</b>: People with motor difficulties and speech impairments often struggle to communicate their needs and views. Augmentative and Alternative Communication (AAC) offers solutions through gestures, body language, or specialized equipment. However, eye gaze and eye signs remain the sole communication method for some individuals. While existing eye-gaze devices leverage deep learning, their pre-calibration techniques can be unreliable and susceptible to lighting conditions. On the other hand, the research into eye sign-based communication is at a very novice stage.</p><p><p><b>Methods</b>: In this research, we propose an eye sign-based communication system that operates on deep learning principles and accepts eye sign patterns from speech-impaired or paraplegic individuals via a standard webcam. The system converts the eye signs into alphabets, words, or sentences and displays the resulting text visually on the screen. In addition, it provides a vocal prompt for the user and the caretaker. It functions effectively in various lighting conditions without requiring calibration and integrates a text prediction function for user convenience. Impact Experiments conducted with participants aged between 18 and 35 years yielded average accuracy rates of 98%, 99%, and 99% for alphabet, word, and sentence formation, respectively. These results demonstrate the system's robustness and potential to significantly benefit individuals with speech impairments.</p>","PeriodicalId":47806,"journal":{"name":"Disability and Rehabilitation-Assistive Technology","volume":" ","pages":"1-22"},"PeriodicalIF":2.2000,"publicationDate":"2025-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Deep learning-based eye sign communication system for people with speech impairments.\",\"authors\":\"Rajesh Kannan Megalingam, Sakthiprasad Kuttankulangara Manoharan, Gokul Riju, Shree Rajesh Raagul Vadivel\",\"doi\":\"10.1080/17483107.2025.2532698\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p><b>Objective</b>: People with motor difficulties and speech impairments often struggle to communicate their needs and views. Augmentative and Alternative Communication (AAC) offers solutions through gestures, body language, or specialized equipment. However, eye gaze and eye signs remain the sole communication method for some individuals. While existing eye-gaze devices leverage deep learning, their pre-calibration techniques can be unreliable and susceptible to lighting conditions. On the other hand, the research into eye sign-based communication is at a very novice stage.</p><p><p><b>Methods</b>: In this research, we propose an eye sign-based communication system that operates on deep learning principles and accepts eye sign patterns from speech-impaired or paraplegic individuals via a standard webcam. The system converts the eye signs into alphabets, words, or sentences and displays the resulting text visually on the screen. In addition, it provides a vocal prompt for the user and the caretaker. It functions effectively in various lighting conditions without requiring calibration and integrates a text prediction function for user convenience. Impact Experiments conducted with participants aged between 18 and 35 years yielded average accuracy rates of 98%, 99%, and 99% for alphabet, word, and sentence formation, respectively. These results demonstrate the system's robustness and potential to significantly benefit individuals with speech impairments.</p>\",\"PeriodicalId\":47806,\"journal\":{\"name\":\"Disability and Rehabilitation-Assistive Technology\",\"volume\":\" \",\"pages\":\"1-22\"},\"PeriodicalIF\":2.2000,\"publicationDate\":\"2025-07-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Disability and Rehabilitation-Assistive Technology\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1080/17483107.2025.2532698\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"REHABILITATION\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Disability and Rehabilitation-Assistive Technology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1080/17483107.2025.2532698","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"REHABILITATION","Score":null,"Total":0}
Deep learning-based eye sign communication system for people with speech impairments.
Objective: People with motor difficulties and speech impairments often struggle to communicate their needs and views. Augmentative and Alternative Communication (AAC) offers solutions through gestures, body language, or specialized equipment. However, eye gaze and eye signs remain the sole communication method for some individuals. While existing eye-gaze devices leverage deep learning, their pre-calibration techniques can be unreliable and susceptible to lighting conditions. On the other hand, the research into eye sign-based communication is at a very novice stage.
Methods: In this research, we propose an eye sign-based communication system that operates on deep learning principles and accepts eye sign patterns from speech-impaired or paraplegic individuals via a standard webcam. The system converts the eye signs into alphabets, words, or sentences and displays the resulting text visually on the screen. In addition, it provides a vocal prompt for the user and the caretaker. It functions effectively in various lighting conditions without requiring calibration and integrates a text prediction function for user convenience. Impact Experiments conducted with participants aged between 18 and 35 years yielded average accuracy rates of 98%, 99%, and 99% for alphabet, word, and sentence formation, respectively. These results demonstrate the system's robustness and potential to significantly benefit individuals with speech impairments.