Ziwei Song, Yuichiro Kinoshita, K. Go, Gangyong Jia
{"title":"Touch Point Prediction for Interactive Public Displays Based on Camera Images","authors":"Ziwei Song, Yuichiro Kinoshita, K. Go, Gangyong Jia","doi":"10.1109/CW52790.2021.00029","DOIUrl":null,"url":null,"abstract":"Feedback latency during the use of interactive displays is an issue currently being considered in the HCI field. Several studies have focused on reducing latency using various approaches. This paper proposes a framework that uses a convolutional neural network to predict user touch points for interactive public displays. The framework predicts user touch events before the finger reaches the display surface to reduce the latency in feedback. As a training dataset, 1,651 tapping actions were collected from 18 participants in front of a display. The training of the convolutional neural network architecture was performed using the collected tapping actions. Validation test results showed that reasonable accuracy could be achieved at 390 ms before touching the display.","PeriodicalId":199618,"journal":{"name":"2021 International Conference on Cyberworlds (CW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Cyberworlds (CW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CW52790.2021.00029","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Feedback latency during the use of interactive displays is an issue currently being considered in the HCI field. Several studies have focused on reducing latency using various approaches. This paper proposes a framework that uses a convolutional neural network to predict user touch points for interactive public displays. The framework predicts user touch events before the finger reaches the display surface to reduce the latency in feedback. As a training dataset, 1,651 tapping actions were collected from 18 participants in front of a display. The training of the convolutional neural network architecture was performed using the collected tapping actions. Validation test results showed that reasonable accuracy could be achieved at 390 ms before touching the display.