{"title":"Synthetic and privacy-preserving visualization of video sensor network outputs","authors":"Carmelo Velardo, Claudia Araimo, J. Dugelay","doi":"10.1109/ICDSC.2011.6042938","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042938","url":null,"abstract":"We propose a complete framework for people tracking in video surveillance networks that preserves privacy by mapping people recorded by a video sensor network in the map of the corresponding surveilled areas. Thanks to this approach, it is possible (1) to synthesize multiple video outputs into a unique picture, and (2) to control privacy by offering the possibility to filter information to viewers. A set of physical attributes, namely soft biometric characteristics, is used to provide the system with tracking capabilities. We demonstrate the feasibility of our approach by showing a real case application scenario.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128859242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Evaluation of similarity measures for appearance-based multi-camera matching","authors":"J. Sherrah, D. Kamenetsky, T. Scoleri","doi":"10.1109/ICDSC.2011.6042930","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042930","url":null,"abstract":"Visually matching people appearing in different camera views is an essential part of multi-camera tracking, camera hand-over and video-based identity search. The problem is made difficult by large variations in the appearance of subjects both within the same camera view and between cameras, as well as across time. Rather than relying on a single appearance-based matching method, a fusion of visual cues is more compelling. In this work 8 different similarity measures were evaluated encompassing shape, colour, texture and biometric information. The evaluation was performed on hand-labelled data from 4 indoor surveillance cameras. Experiments examined the accuracy of the similarity measures. Results revealed that matching accuracy is good when tracks come from the same camera, but poor when they come from different cameras. Although different measures performed best in different situations, the colour-based measure produced the best results overall.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127389362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Accelerating Viola-Jones face detection for embedded and SoC environments","authors":"Laurentiu Acasandrei, A. Barriga","doi":"10.1109/ICDSC.2011.6042932","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042932","url":null,"abstract":"In this communication a speed optimized implementation of Viola-Jones Face Detection Algorithm based on the baseline OpenCV face detection application is presented. The baseline OpenCV face detection application is analyzed. Then the necessary modifications and improvements are described in order to accelerate the execution speed in an embedded or SoC (System-on-Chip) environments.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130456355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"PhD forum: Volumetric 3D reconstruction without planar ground assumption","authors":"H. Aliakbarpour, J. Dias","doi":"10.1109/ICDSC.2011.6042942","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042942","url":null,"abstract":"This paper proposes a framework to perform volumetric 3D reconstruction using a camera network. A network of cameras observes a scene and each camera is rigidly coupled with an Inertial Sensor (IS). The 3D orientation provided by IS is used firstly for definition of a virtual camera network whose axis are aligned to the earth cardinal directions. Then a set of virtual planes are defined for the sake of 3D reconstruction with no planar ground assumption and just by using 3D orientation data provided by IS. A GPU-based implementation of the proposed method is provided to demonstrate the promising results.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131961457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
N. Deligiannis, M. Jacobs, Frederik Verbist, Jürgen Slowack, J. Barbarien, R. Walle, P. Schelkens, A. Munteanu
{"title":"Efficient hash-driven Wyner-Ziv video coding for visual sensors","authors":"N. Deligiannis, M. Jacobs, Frederik Verbist, Jürgen Slowack, J. Barbarien, R. Walle, P. Schelkens, A. Munteanu","doi":"10.1109/ICDSC.2011.6042921","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042921","url":null,"abstract":"Recent advances in wireless visual sensor technology, have been calling for innovative architectures realizing efficient video coding under stringent processing and energy restrictions. Driven by profound findings in network information theory, Wyner-Ziv video coding constitutes a suitable paradigm for video sensor networks. This work presents a novel hash-driven Wyner-Ziv video coding architecture for visual sensors, which coarsely encodes a low resolution version of each Wyner-Ziv frame to facilitate accurate motion-compensated prediction at the decoder. The proposed method for side-information generation comprises hash-based multi-hypothesis pixel-based prediction. Once critical Wyner-Ziv information is decoded, the derived dense motion field is further enhanced. Experimental validation illustrates that the proposed hash-driven codec achieves significant compression gains with respect to state-of-the-art Wyner-Ziv video coding, even under demanding conditions.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122190926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fusion of non-visual modalities into the Probabilistic Occupancy Map framework for person localization","authors":"Rok Mandeljc, J. Pers, M. Kristan, S. Kovacic","doi":"10.1109/ICDSC.2011.6042937","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042937","url":null,"abstract":"In this paper we investigate the possibilities for fusion of non-visual sensor modalities into state-of-the-art vision-based framework for person detection and localization, the Probabilistic Occupancy Map (POM), with the aim of improving the frame-by-frame localization results in a realistic (cluttered) indoor environment. We point out the aspects that need to be considered when fusing non-visual sensor information into POM and provide a mathematical model for it. We demonstrate the proposed fusion method on the example of multi-camera and radio-based person localization setup. The performance of both systems is evaluated, showing their strengths and weaknesses. We show that localization results may be significantly improved by fusing the information from the radio-based system into the camera-based POM framework using the proposed model.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116400980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Human tracking by adaptive Kalman filtering and multiple kernels tracking with projected gradients","authors":"Chun-Te Chu, Jenq-Neng Hwang, Shen-Zheng Wang, Yi-Yuan Chen","doi":"10.1109/ICDSC.2011.6042939","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042939","url":null,"abstract":"Kernel based trackers have been proven to be a promising approach in video object tracking. The use of single kernel often suffers from occlusion since the visual information is not sufficient for kernel usage. Hence, multiple inter-related kernels have been utilized for tracking in complicated scenarios. This paper embeds the multiple kernels tracking into a Kalman filtering-based tracking system, which uses Kalman prediction as the initial position for the multiple kernels tracking, and applies the result of the latter as the measurement to the Kalman update. The state transition and noise covariance matrices used in Kalman filter are also dynamically updated by the output of multiple kernels tracking. Several simulation results have been done to show the robustness of the proposed system which can successfully track all the video objects under occlusion.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127139202","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Demo: Spatio-temporal template matching for ball detection","authors":"K. Kumar, Pascaline Parisot, C. Vleeschouwer","doi":"10.1109/ICDSC.2011.6042940","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042940","url":null,"abstract":"This paper considers the detection of ball in a basketball game covered by multiple loosely synchronized cameras. First, plausible ball candidates are detected on the nodes of a 3D grid defined around the basket. This is done by correlating independently in each view the spatial template of the ball with a precomputed foreground mask. Efficient implementation of this step relies on integral images. Afterwards, false positives are filtered out based on a temporal analysis of the ball trajectory. This analysis builds on the Random Sample Consensus (RANSAC) method, with a ballistic trajectory model. The integrated approach is demonstrated on a real-life dataset, and appears to be both effective and efficient.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126379686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Demo: A distributed virtual vision simulator","authors":"Wiktor Starzyk, Adam Domurad, F. Qureshi","doi":"10.1109/ICDSC.2011.6042945","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042945","url":null,"abstract":"Realistic virtual worlds can serve as laboratories for carrying out camera networks research. This unorthodox “Virtual Vision” paradigm advocates developing visually and behaviorally realistic 3D environments to serve the needs of computer vision. Our work on high-level coordination and control in camera networks is a testament to the suitability of virtual vision paradigm for camera networks research. The prerequisite for carrying out virtual vision research is a virtual vision simulator capable of generating synthetic imagery from simulated real-life scenes. We present a distributed, customizable virtual vision simulator capable of simulating pedestrian traffic in a variety of 3D environments. Virtual cameras deployed in this synthetic environment generate synthetic imagery — boasting realistic lighting effects, shadows, etc. — using the state-of-the-art computer graphics techniques. The synthetic imagery is fed into a “real-world” vision pipeline that performs visual analysis — e.g., blob detection and tracking, facial detection, etc. — and returns the results of this analysis to our simulated cameras for subsequent higher level processing. It is important to bear in mind that our vision pipeline is designed to handle real world imagery without any modifications. Consequently, it closely mimics the performance of a vision pipeline that one might deploy on physical cameras. Our virtual vision simulator is realized as a collection of modules that communicate with each other over the network. Consequently, we can deploy our simulator over a network of computers, allowing us to simulate much larger networks and much more complex scenes then is otherwise possible.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"106 10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130661616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yan Yang, F. Dadgostar, Conrad Sanderson, B. Lovell
{"title":"Summarisation of surveillance videos by key-frame selection","authors":"Yan Yang, F. Dadgostar, Conrad Sanderson, B. Lovell","doi":"10.1109/ICDSC.2011.6042925","DOIUrl":"https://doi.org/10.1109/ICDSC.2011.6042925","url":null,"abstract":"We propose two novel techniques for automatic summarisation of lengthy surveillance videos, based on selection of frames containing scenes most informative for rapid perusal and interpretation by humans. In contrast to other video summarisation methods, the proposed methods explicitly focus on foreground objects, via edge histogram descriptor and a localised foreground information quantity (entropy) measurement. Frames are iteratively pruned until a preset summarisation rate is reached. Experiments on the publicly available CAVIAR dataset, as well as our own dataset focused on people walking through natural choke points (such as doors), suggest that the proposed method obtains considerably better results than methods based on optical flow, entropy differences and colour spatial distribution characteristics.","PeriodicalId":385052,"journal":{"name":"2011 Fifth ACM/IEEE International Conference on Distributed Smart Cameras","volume":"2003 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127695173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}