Mustafa Shdaifat, S. S. Bukhari, Takumi Toyama, A. Dengel
{"title":"Robust object recognition in wearable eye tracking system","authors":"Mustafa Shdaifat, S. S. Bukhari, Takumi Toyama, A. Dengel","doi":"10.1109/ACPR.2015.7486583","DOIUrl":null,"url":null,"abstract":"Object recognition is a versatile capability. Automatic guided tours and augmented reality are just two examples. Humans seem to do it subconsciously - unaware of the extensive processing required for it - while it is a complex task for machines. Methods based on SIFT features have proven to be robust for recognition. However, a prior detection step is required to limit confusion, caused by, e.g., scene clutter. We present an attention-guided method that offloads this to humans through eye tracking. Gaze data is used to extract candidate patches to recognize afterwards. It improves upon previous work by automatically selecting the dynamic size of said patch, instead of fixed large local region. Therefore increasing robustness and independence compared to fixed window size technique.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"32 4","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACPR.2015.7486583","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Object recognition is a versatile capability. Automatic guided tours and augmented reality are just two examples. Humans seem to do it subconsciously - unaware of the extensive processing required for it - while it is a complex task for machines. Methods based on SIFT features have proven to be robust for recognition. However, a prior detection step is required to limit confusion, caused by, e.g., scene clutter. We present an attention-guided method that offloads this to humans through eye tracking. Gaze data is used to extract candidate patches to recognize afterwards. It improves upon previous work by automatically selecting the dynamic size of said patch, instead of fixed large local region. Therefore increasing robustness and independence compared to fixed window size technique.