Machine Vision Intelligent Travel Aid for the Visually Impaired (ITAVI) in Developing Countries*

S. Atiba, Sarah Funmilola Moses, M. Lakoju, F. A. Semire, R. Aldmour
{"title":"Machine Vision Intelligent Travel Aid for the Visually Impaired (ITAVI) in Developing Countries*","authors":"S. Atiba, Sarah Funmilola Moses, M. Lakoju, F. A. Semire, R. Aldmour","doi":"10.1109/DeSE51703.2020.9450744","DOIUrl":null,"url":null,"abstract":"The visually impaired have little or no effective visual sensory input and have to rely on external assistance for navigation. Several electronic travel aids have been developed to aid independent navigation of the visually impaired, however they are not without limitations and dependence on third-parties. This paper describes the design and implementation of an Intelligent Travel Aid for the Visually Impaired, it combines the detection and recognition of objects in real-time with audio feedback to provide aid to the visually impaired users. This assistive device uses machine vision for object recognition detection, a camera for capturing object images respectively and a speaker all collectively form the core of the system. The system notifies users of obstacles and objects via synthesized speech. Using a quantized MobileNet based Single Shot multibox object detection model pre-trained on the Common Objects in Context dataset, the device was able to detect objects/obstacles, as well as determine the relative position and approximate distance. The device, when tested, was found to achieve real time performance of up to 70.56 frames per second for detections. Audio feedback was also achieved using the eSpeak Text to Speech engine to provide real time voice instructions to the user. All algorithms were implemented using Python language. The device is user friendly, allowing the visually impaired to enjoy easier navigation. However, other features such as extra object classes as well as language variety could be added in order to boost the robustness of the device.","PeriodicalId":124051,"journal":{"name":"2020 13th International Conference on Developments in eSystems Engineering (DeSE)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 13th International Conference on Developments in eSystems Engineering (DeSE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DeSE51703.2020.9450744","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

The visually impaired have little or no effective visual sensory input and have to rely on external assistance for navigation. Several electronic travel aids have been developed to aid independent navigation of the visually impaired, however they are not without limitations and dependence on third-parties. This paper describes the design and implementation of an Intelligent Travel Aid for the Visually Impaired, it combines the detection and recognition of objects in real-time with audio feedback to provide aid to the visually impaired users. This assistive device uses machine vision for object recognition detection, a camera for capturing object images respectively and a speaker all collectively form the core of the system. The system notifies users of obstacles and objects via synthesized speech. Using a quantized MobileNet based Single Shot multibox object detection model pre-trained on the Common Objects in Context dataset, the device was able to detect objects/obstacles, as well as determine the relative position and approximate distance. The device, when tested, was found to achieve real time performance of up to 70.56 frames per second for detections. Audio feedback was also achieved using the eSpeak Text to Speech engine to provide real time voice instructions to the user. All algorithms were implemented using Python language. The device is user friendly, allowing the visually impaired to enjoy easier navigation. However, other features such as extra object classes as well as language variety could be added in order to boost the robustness of the device.
发展中国家为视障人士提供的机器视觉智能旅行辅助工具(ITAVI) *
视障人士很少或没有有效的视觉感官输入,必须依靠外部辅助来导航。已经开发了一些电子旅行辅助工具来帮助视障人士独立导航,但是它们并非没有限制和依赖第三方。本文介绍了一种视障人士智能助行器的设计与实现,它将物体的实时检测与识别与音频反馈相结合,为视障人士提供辅助。该辅助装置采用机器视觉进行物体识别检测,相机分别捕捉物体图像,扬声器共同构成系统的核心。该系统通过合成语音通知用户障碍物和物体。使用量化的基于MobileNet的单镜头多盒物体检测模型,该设备能够检测物体/障碍物,并确定相对位置和近似距离。经过测试,发现该设备的实时检测性能高达每秒70.56帧。使用eSpeak文本到语音引擎也实现了音频反馈,为用户提供实时语音指令。所有算法均使用Python语言实现。该设备使用方便,让视障人士更容易导航。然而,为了增强设备的健壮性,可以添加其他功能,如额外的对象类和语言多样性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信