{"title":"基于深度学习的视障可穿戴视觉-听觉感官替代系统","authors":"Zifeng Wang, Heng Li, Jianping Chen, X. Chai, Zhenzhen Zhai","doi":"10.1109/dsins54396.2021.9670599","DOIUrl":null,"url":null,"abstract":"Visual impairment has caused serious influence on the human being and society. Due to more sensitive hearing and touch, for visually impaired people, it is an available solution to improve their quality of live, work and study by Sensory Substitution Devices (SSDs) which transfer visual information to audio or touch. In this paper, we proposed a wearable, vision-to-audio sensory substitution system with scene-perception-based deep learning to help the visually impaired users recognize and locate normal objects in the environment. The system consists of a wireless camera module, a Bluetooth speech feedback module with a microphone, and an Android mobile phone with a customized application. The camera module captures images from the scene and sends them to the application of Android mobile phone. The Bluetooth speech feedback module sends speech commands to application and broadcasts speech guidance to visually impaired users. The application based on Android platform loads speech recognition and object detection models. The system has been proved to provide an effective way to help the visually impaired people recognize and locate objects.","PeriodicalId":243724,"journal":{"name":"2021 International Conference on Digital Society and Intelligent Systems (DSInS)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"A Wearable Vision-To-Audio Sensory Substitution System Based on Deep Learning for the Visually Impaired\",\"authors\":\"Zifeng Wang, Heng Li, Jianping Chen, X. Chai, Zhenzhen Zhai\",\"doi\":\"10.1109/dsins54396.2021.9670599\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Visual impairment has caused serious influence on the human being and society. Due to more sensitive hearing and touch, for visually impaired people, it is an available solution to improve their quality of live, work and study by Sensory Substitution Devices (SSDs) which transfer visual information to audio or touch. In this paper, we proposed a wearable, vision-to-audio sensory substitution system with scene-perception-based deep learning to help the visually impaired users recognize and locate normal objects in the environment. The system consists of a wireless camera module, a Bluetooth speech feedback module with a microphone, and an Android mobile phone with a customized application. The camera module captures images from the scene and sends them to the application of Android mobile phone. The Bluetooth speech feedback module sends speech commands to application and broadcasts speech guidance to visually impaired users. The application based on Android platform loads speech recognition and object detection models. The system has been proved to provide an effective way to help the visually impaired people recognize and locate objects.\",\"PeriodicalId\":243724,\"journal\":{\"name\":\"2021 International Conference on Digital Society and Intelligent Systems (DSInS)\",\"volume\":\"5 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Conference on Digital Society and Intelligent Systems (DSInS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/dsins54396.2021.9670599\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Digital Society and Intelligent Systems (DSInS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/dsins54396.2021.9670599","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Wearable Vision-To-Audio Sensory Substitution System Based on Deep Learning for the Visually Impaired
Visual impairment has caused serious influence on the human being and society. Due to more sensitive hearing and touch, for visually impaired people, it is an available solution to improve their quality of live, work and study by Sensory Substitution Devices (SSDs) which transfer visual information to audio or touch. In this paper, we proposed a wearable, vision-to-audio sensory substitution system with scene-perception-based deep learning to help the visually impaired users recognize and locate normal objects in the environment. The system consists of a wireless camera module, a Bluetooth speech feedback module with a microphone, and an Android mobile phone with a customized application. The camera module captures images from the scene and sends them to the application of Android mobile phone. The Bluetooth speech feedback module sends speech commands to application and broadcasts speech guidance to visually impaired users. The application based on Android platform loads speech recognition and object detection models. The system has been proved to provide an effective way to help the visually impaired people recognize and locate objects.