S. Borkowski, J. Crowley, Julien Letessier, F. Bérard
{"title":"User-Centric Design of a Vision System for Interactive Applications","authors":"S. Borkowski, J. Crowley, Julien Letessier, F. Bérard","doi":"10.1109/ICVS.2006.61","DOIUrl":"https://doi.org/10.1109/ICVS.2006.61","url":null,"abstract":"Despite great promise of vision-based user interfaces, commercial employment of such systems remains marginal. Most vision-based interactive systems are one-time, \"proof of concept\" prototypes that demonstrate the interest of a particular image treatment applied to interaction. In general, vision systems require parameter tuning, both during setup and at runtime, and are thus difficult to handle by nonexperts in computer vision. In this paper, we present a pragmatic, developer-centric, service-oriented framework for the construction of vision-based interactive systems. Our framework is designed to allow developers unfamiliar with vision to use computer vision as an interaction modality. To achieve this goal, we address specific developer- and interaction-centric requirements during the design of our system. We validate our approach with an implementation of standard GUI widgets (buttons and sliders) based on computer vision.","PeriodicalId":189284,"journal":{"name":"Fourth IEEE International Conference on Computer Vision Systems (ICVS'06)","volume":"281 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116585299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-camera video resolution enhancement by fusion of spatial disparity and temporal motion fields","authors":"Daniel Hazen, R. Puri, K. Ramchandran","doi":"10.1109/ICVS.2006.43","DOIUrl":"https://doi.org/10.1109/ICVS.2006.43","url":null,"abstract":"We consider the problem of spatio-temporal resolution enhancement of a dynamic video scene by synergistically combining information from multiple video sequences corresponding to the different views of the scene. While prior work has focused on spatio-temporal superresolution by primarily exploiting the blurring correlation induced by the scene imaging process, in this work we exploit another and perhaps a stronger form of correlation the intrinsic correlation present in the dynamic scene. We use the knowledge base from the areas of multi-view computer vision, video coding and signal/image processing to efficiently model the dynamic video scene by exploiting inter-camera (space) as well as intra-camera correlations (time). We present simulation results corresponding to a lowframe rate two-camera stereo video configuration where the sequences are combined to generate higher frame-rate sequences. Our proposed framework has direct application to problems such as multi-view video surveillance and relates closely to problems such as multi-view video compression.","PeriodicalId":189284,"journal":{"name":"Fourth IEEE International Conference on Computer Vision Systems (ICVS'06)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130032630","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Evaluation and Knowledge Representation Formalisms to Improve Video Understanding","authors":"B. Georis, Magale Maziere, F. Brémond","doi":"10.1109/ICVS.2006.23","DOIUrl":"https://doi.org/10.1109/ICVS.2006.23","url":null,"abstract":"This article presents a methodology to build efficient real-time semantic video understanding systems addressing real world problems. In our case, semantic video under- standing consists in the recognition of predefined scenario models in a given application domain starting from a pixel analysis up to a symbolic description of what is happening in the scene viewed by cameras. This methodology proposes to use evaluation to acquire knowledge of programs and to represent this knowledge with appropriate formalisms. First, to obtain efficiency, a formalism enables to model video processing programs and their associated parameter adaptation rules. These rules are written by experts after performing a technical evaluation. Second, a scenario for- malism enables experts to model their needs and to easily refine their scenario models to adapt them to real-life situa- tions. This refinement is performed with an end-user evalu- ation. This second part ensures that systems match end-user expectations. Results are reported for scenario recognition performances on real video sequences taken from a bank agency monitoring application.","PeriodicalId":189284,"journal":{"name":"Fourth IEEE International Conference on Computer Vision Systems (ICVS'06)","volume":"136 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131777523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hand-Gesture Based Text Input for Wearable Computers","authors":"Yang Liu, Xiabi Liu, Yunde Jia","doi":"10.1109/ICVS.2006.34","DOIUrl":"https://doi.org/10.1109/ICVS.2006.34","url":null,"abstract":"This paper proposes a novel text input method based on hand gestures for wearable computers. A character is firstly written by a fingertip and then recognized through a B-splines based character recognition method. The writing procedure is controlled by hand gestures including hand tracking, gesture recognition and fingertip positioning which are performed by an extended CONDENSATION algorithm. We have integrated the proposed text input method into a wearable vision system developed in our lab and tested the resulting text input system on Graffiti 2 alphabet. The experimental result shows that our method is promising for natural text input in wearable computers.","PeriodicalId":189284,"journal":{"name":"Fourth IEEE International Conference on Computer Vision Systems (ICVS'06)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116964993","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mohamed E. Hussein, W. Abd-Almageed, Yang Ran, L. Davis
{"title":"Real-Time Human Detection, Tracking, and Verification in Uncontrolled Camera Motion Environments","authors":"Mohamed E. Hussein, W. Abd-Almageed, Yang Ran, L. Davis","doi":"10.1109/ICVS.2006.52","DOIUrl":"https://doi.org/10.1109/ICVS.2006.52","url":null,"abstract":"In environments where a camera is installed on a freely moving platform, e.g. a vehicle or a robot, object detection and tracking becomes much more difficult. In this paper, we presents a real time system for human detection, tracking, and verification in such challenging environments. To deliver a robust performance, the system integrates several computer vision algorithms to perform its function: a human detection algorithm, an object tracking algorithm, and a motion analysis algorithm. To utilize the available computing resources to the maximum possible extent, each of the system components is designed to work in a separate thread that communicates with the other threads through shared data structures. The focus of this paper is more on the implementation issues than on the algorithmic issues of the system. Object oriented design was adopted to abstract algorithmic details away from the system structure.","PeriodicalId":189284,"journal":{"name":"Fourth IEEE International Conference on Computer Vision Systems (ICVS'06)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121493311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}