Fanghao Song, Zhongen Li, Brian C. Clark, Dustin R. Grooms, Chang Liu
{"title":"Camera-Based Indoor Navigation in Known Environments with ORB for People with Visual Impairment","authors":"Fanghao Song, Zhongen Li, Brian C. Clark, Dustin R. Grooms, Chang Liu","doi":"10.1109/GHTC46280.2020.9342876","DOIUrl":null,"url":null,"abstract":"Indoor navigation is a hard problem for those with visual impairment or blindness. This paper presents a real-time camera-based indoor navigation application for the blind and visually impaired in known environments. Contrary to other systems for similar purposes, our system executes all computation locally on mobile devices and does not need any infrastructure-based sensor and signal support. The use of the system is divided into two stages: an offline stage, in which a sighted person helps with landmark identification and route planning, and an online stage, in which a visually impaired person can navigate with voice prompts from the app. Predefined landmark images and navigation information are stored in a database in the offline stage. Navigation information consists of the landmark position, the turning angle, the turning direction, and the distance to the next landmark. In the online stage, navigation information is retrieved from the current landmark image by successfully matching video frames in real time. A similarity score is calculated by ORB using Hamming distances. Voice feedback is provided for users with a text-to-speech function. Our system minimizes the error caused by users’ manipulation and improves stability and accuracy by updating similarity scores continuously.","PeriodicalId":314837,"journal":{"name":"2020 IEEE Global Humanitarian Technology Conference (GHTC)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE Global Humanitarian Technology Conference (GHTC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/GHTC46280.2020.9342876","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Indoor navigation is a hard problem for those with visual impairment or blindness. This paper presents a real-time camera-based indoor navigation application for the blind and visually impaired in known environments. Contrary to other systems for similar purposes, our system executes all computation locally on mobile devices and does not need any infrastructure-based sensor and signal support. The use of the system is divided into two stages: an offline stage, in which a sighted person helps with landmark identification and route planning, and an online stage, in which a visually impaired person can navigate with voice prompts from the app. Predefined landmark images and navigation information are stored in a database in the offline stage. Navigation information consists of the landmark position, the turning angle, the turning direction, and the distance to the next landmark. In the online stage, navigation information is retrieved from the current landmark image by successfully matching video frames in real time. A similarity score is calculated by ORB using Hamming distances. Voice feedback is provided for users with a text-to-speech function. Our system minimizes the error caused by users’ manipulation and improves stability and accuracy by updating similarity scores continuously.