Reham H. Elnabawy, Slim Abdennadher, O. Hellwich, S. Eldawlatly
{"title":"基于yolo的视觉假体对象简化方法","authors":"Reham H. Elnabawy, Slim Abdennadher, O. Hellwich, S. Eldawlatly","doi":"10.1109/CBMS55023.2022.00039","DOIUrl":null,"url":null,"abstract":"Visual prostheses have been introduced to partially restore vision to the blind via visual pathway stimulation. Despite their success, some challenges have been reported by the implanted patients. One of those challenges is the difficulty of object recognition due to the low resolution of the images perceived through these devices. In this paper, a deep learning-based approach combined with image pre-processing is proposed to allow visual prostheses' users to recognize objects in a given scene. The approach simplifies the objects in the scene by displaying the objects in clip art form to enhance object recognition. These clip art images are generated by, first, identifying the objects in the scene using the You Only Look Once (YOLO) deep neural network. The clip art corresponding to each identified object is then retrieved via Google Images. Three experiments were conducted to measure the success of the proposed approach using simulated prosthetic vision. Our results reveal a remarkable decrease in the recognition time, increase in the recognition accuracy and confidence level when using the clip art representation as opposed to using the actual images of the objects. These results demonstrate the utility of object simplification in enhancing the perception of images in prosthetic vision.","PeriodicalId":218475,"journal":{"name":"2022 IEEE 35th International Symposium on Computer-Based Medical Systems (CBMS)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"A YOLO-based Object Simplification Approach for Visual Prostheses\",\"authors\":\"Reham H. Elnabawy, Slim Abdennadher, O. Hellwich, S. Eldawlatly\",\"doi\":\"10.1109/CBMS55023.2022.00039\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Visual prostheses have been introduced to partially restore vision to the blind via visual pathway stimulation. Despite their success, some challenges have been reported by the implanted patients. One of those challenges is the difficulty of object recognition due to the low resolution of the images perceived through these devices. In this paper, a deep learning-based approach combined with image pre-processing is proposed to allow visual prostheses' users to recognize objects in a given scene. The approach simplifies the objects in the scene by displaying the objects in clip art form to enhance object recognition. These clip art images are generated by, first, identifying the objects in the scene using the You Only Look Once (YOLO) deep neural network. The clip art corresponding to each identified object is then retrieved via Google Images. Three experiments were conducted to measure the success of the proposed approach using simulated prosthetic vision. Our results reveal a remarkable decrease in the recognition time, increase in the recognition accuracy and confidence level when using the clip art representation as opposed to using the actual images of the objects. These results demonstrate the utility of object simplification in enhancing the perception of images in prosthetic vision.\",\"PeriodicalId\":218475,\"journal\":{\"name\":\"2022 IEEE 35th International Symposium on Computer-Based Medical Systems (CBMS)\",\"volume\":\"4 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 35th International Symposium on Computer-Based Medical Systems (CBMS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CBMS55023.2022.00039\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 35th International Symposium on Computer-Based Medical Systems (CBMS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CBMS55023.2022.00039","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
摘要
视觉假体通过视觉通路刺激来部分恢复盲人的视力。尽管他们取得了成功,但植入患者也报告了一些挑战。其中一个挑战是物体识别的困难,因为通过这些设备感知的图像分辨率很低。本文提出了一种基于深度学习与图像预处理相结合的方法,使视觉假体的用户能够识别给定场景中的物体。该方法通过将对象以剪贴画的形式显示来简化场景中的对象,从而增强对象的识别能力。首先,这些剪贴画图像是通过使用You Only Look Once (YOLO)深度神经网络识别场景中的对象生成的。然后通过Google Images检索每个识别对象对应的剪贴画。通过模拟假体视觉进行了三个实验来衡量所提出方法的成功。我们的结果显示,与使用对象的实际图像相比,使用剪贴画表示显著减少了识别时间,提高了识别精度和置信度。这些结果证明了物体简化在增强假肢视觉图像感知方面的效用。
A YOLO-based Object Simplification Approach for Visual Prostheses
Visual prostheses have been introduced to partially restore vision to the blind via visual pathway stimulation. Despite their success, some challenges have been reported by the implanted patients. One of those challenges is the difficulty of object recognition due to the low resolution of the images perceived through these devices. In this paper, a deep learning-based approach combined with image pre-processing is proposed to allow visual prostheses' users to recognize objects in a given scene. The approach simplifies the objects in the scene by displaying the objects in clip art form to enhance object recognition. These clip art images are generated by, first, identifying the objects in the scene using the You Only Look Once (YOLO) deep neural network. The clip art corresponding to each identified object is then retrieved via Google Images. Three experiments were conducted to measure the success of the proposed approach using simulated prosthetic vision. Our results reveal a remarkable decrease in the recognition time, increase in the recognition accuracy and confidence level when using the clip art representation as opposed to using the actual images of the objects. These results demonstrate the utility of object simplification in enhancing the perception of images in prosthetic vision.