D. Looney, N. Rehman, D. Mandic, Tomasz M. Rutkowski, A. Heidenreich, Dagmar Beyer
{"title":"Conditioning multimodal information for smart environments","authors":"D. Looney, N. Rehman, D. Mandic, Tomasz M. Rutkowski, A. Heidenreich, Dagmar Beyer","doi":"10.1109/ICDSC.2009.5289373","DOIUrl":"https://doi.org/10.1109/ICDSC.2009.5289373","url":null,"abstract":"This study aims at providing signal processing solutions for the conditioning of multimodal information in audio-aided smart camera environments. A novel approach is introduced for processing audio and video within a unified ‘data fusion via fission’ framework. This is achieved using empirical mode decomposition (EMD), a fully data-driven algorithm which facilitates analysis at multiple time-frequency scales. Its adaptive nature makes it suitable for processing real-world data and allows, for example, signal conditioning (denoising, illumination invariant video) and robust feature extraction. Furthermore, complex extension of the EMD algorithm are used to quantify shared dynamics between the conditioned modalities facilitating multimodal fusion. The proposed collaborative approach is used to model human-human interaction.","PeriodicalId":324810,"journal":{"name":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133935493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
N. Anantrasirichai, D. Agrafiotis, M. Ouaret, F. Dufaux, T. Ebrahimi
{"title":"Performance evaluation of two state of the art DVC codecs","authors":"N. Anantrasirichai, D. Agrafiotis, M. Ouaret, F. Dufaux, T. Ebrahimi","doi":"10.1109/ICDSC.2009.5289372","DOIUrl":"https://doi.org/10.1109/ICDSC.2009.5289372","url":null,"abstract":"The performance of existing DVC codecs is still lacking relative to that of H.264 and work is being carried out in order to close this gap. The authors of this paper have been and still are involved in the development of two DVC codecs respectively, the performance of which is compared herein. The aim is to identify strengths and weaknesses of the two codecs that can be exploited / addressed in order to improve the achieved performance relative to H.264.","PeriodicalId":324810,"journal":{"name":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117132374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hojin Kee, S. Bhattacharyya, N. Petersen, Jacob Kornerup
{"title":"Resource-efficient acceleration of 2-dimensional Fast Fourier Transform computations on FPGAs","authors":"Hojin Kee, S. Bhattacharyya, N. Petersen, Jacob Kornerup","doi":"10.1109/ICDSC.2009.5289356","DOIUrl":"https://doi.org/10.1109/ICDSC.2009.5289356","url":null,"abstract":"The 2-dimensional (2D) Fast Fourier Transform (FFT) is a fundamental, computationally intensive function that is of broad relevance to distributed smart camera systems. In this paper, we develop a systematic method for improving the throughput of 2D-FFT implementations on field-programmable gate arrays (FPGAs). Our method is based on a novel loop unrolling technique for FFT implementation, which is extended from our recent work on FPGA architectures for 1D-FFT implementation [1]. This unrolling technique deploys multiple processing units within a single 1D-FFT core to achieve efficient configurations of data parallelism while minimizing memory space requirements, and FPGA slice consumption. Furthermore, using our techniques for parallel processing within individual 1DFFT cores, the number of input/output (I/O) ports within a given 1D-FFT core is limited to one input port and one output port. In contrast, previous 2D-FFT design approaches require multiple I/O pairs with multiple FFT cores. This streamlining of 1D-FFT interfaces makes it possible to avoid complex interconnection networks and associated scheduling logic for connecting multiple I/O ports from 1D-FFT cores to the I/O channel of external memory devices. Hence, our proposed unrolling technique maximizes the ratio of the achieved throughput to the consumed FPGA resources under pre-defined constraints on I/O channel bandwidth. To provide generality, our framework for 2D-FFT implementation can be efficiently parameterized in terms of key design parameters such as the transform size and I/O data word length.","PeriodicalId":324810,"journal":{"name":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","volume":"284 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116190876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Semantic browsing of video surveillance databases through Online Generic Indexing","authors":"Denis Marraud, Benjamin Cepas, Livier Reithler","doi":"10.1109/ICDSC.2009.5289366","DOIUrl":"https://doi.org/10.1109/ICDSC.2009.5289366","url":null,"abstract":"This paper gives a thorough overview of EADS UrbanVIEW indexing and mining platform aimed at providing police forces and security officers with advanced tools to efficiently browse large video surveillance databases for investigation purposes. A scalable indexing architecture that works indifferently with smart or classical camera networks as well as for real-time or a posteriori indexing has been designed and implemented. We introduce the concept of Online Generic Indexing Strategy (OGIS) aimed at systematically enriching each video stream with real-time extracted generic metadata allowing to dramatically decrease post-event investigation time. The indexing strategy relies on the systematic detection, tracking and characterization of all observed moving objects. Semantic and non semantic metadata produced by embedded or distributed video analytics modules can be used either to browse the distributed video databases or as inputs to higher level characterization modules (object identification, multi-camera back-tracking, event recognition…). Once a first observation of an object of interest has been found, it can be forward and backward tracked thanks to an interactive multi-stream player taking into account the multi-camera context. Our platform has been assessed on the NGSIM and I-LIDS datasets which consist of real heavy traffic images, showing both high recall and high detection rates in its retrieval capabilities.","PeriodicalId":324810,"journal":{"name":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126478557","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A distributed camera network architecture supporting video adaptation","authors":"Razib Iqbal, S. Ratti, S. Shirmohammadi","doi":"10.1109/ICDSC.2009.5289376","DOIUrl":"https://doi.org/10.1109/ICDSC.2009.5289376","url":null,"abstract":"Sparked by the desire to secure human presence in high risk areas, this paper proposes a distributed camera network architecture. This architecture is intended as a robust platform for video surveillance systems. In addition, the proposed system has the ability to collect/capture video streams from cameras at a point-of-interest and deliver a consolidated version of the video to a receiver. To achieve this, video sources are able to discover nodes, adapt content, and transmit adapted video streams to a receiver. Our framework takes into account different issues related to the decentralized Peer-to-Peer paradigm such as peers' unreliability, as well as pragmatic aspects like receivers' heterogeneity. Adaptation ensures that video quality is proportional to bandwidth availability as well as the receiving devices' resources. Simulation is used to manifest that the framework is robust, reliable and suitable for multiparticipant real-time collaboration and real-life deployment.","PeriodicalId":324810,"journal":{"name":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125510690","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pierre Clarot, E. Ermis, Pierre-Marc Jodoin, Venkatesh Saligrama
{"title":"Unsupervised camera network structure estimation based on activity","authors":"Pierre Clarot, E. Ermis, Pierre-Marc Jodoin, Venkatesh Saligrama","doi":"10.1109/ICDSC.2009.5289362","DOIUrl":"https://doi.org/10.1109/ICDSC.2009.5289362","url":null,"abstract":"In this paper we consider the problem of unsupervised topology reconstruction in uncalibrated visual sensor networks. We assume that a number of video cameras observe a common scene from arbitrary and unknown locations, orientations and zoom levels, and show that the extrinsic and calibration matrices, fundamental and essential matrices, the homography matrix, and the physical configuration of the cameras with respect to each other can be estimated in an unsupervised manner. Our method relies on the similarity of activity patterns observed at various locations, and an unsupervised matching method based on these activity patterns. The proposed method works in cases with cameras having significantly different orientations and zoom levels, where many of the existing methods cannot be applied. We explain how to extend the method to a multicamera case where more than two cameras are involved. We present both qualitative and quantitative results of our estimates, and conclude that this method can be applied in wide area surveillance applications in which the deployed systems need to be flexible and scalable, and where calibration can be a major challenge.","PeriodicalId":324810,"journal":{"name":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","volume":"72 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124677408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"PhD forum: Multiple camera management using wide base-line matching","authors":"R. Lakemond, C. Fookes, S. Sridharan","doi":"10.1109/ICDSC.2009.5289393","DOIUrl":"https://doi.org/10.1109/ICDSC.2009.5289393","url":null,"abstract":"Manual calibration of large and dynamic networks of cameras is labor intensive and time consuming. This is a strong motivator for the development of automatic calibration methods. Automatic calibration relies on the ability to find correspondences between multiple views of the same scene. If the cameras are sparsely placed, this can be a very difficult task. This PhD project focuses on the further development of uncalibrated wide baseline matching techniques.","PeriodicalId":324810,"journal":{"name":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117021414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Wided Miled, Thomas Maugey, Marco Cagnazzo, B. Pesquet-Popescu
{"title":"Image interpolation with dense disparity estimation in multiview distributed video coding","authors":"Wided Miled, Thomas Maugey, Marco Cagnazzo, B. Pesquet-Popescu","doi":"10.1109/ICDSC.2009.5289344","DOIUrl":"https://doi.org/10.1109/ICDSC.2009.5289344","url":null,"abstract":"This paper deals with the side information (SI) generation problem in multiview distributed video coding (MDVC). For inter-view interpolation, we propose a novel dense disparity estimation (DE) approach combined with a popular distributed video coding (DVC) technique. As disparity vectors are computed at the decoder side, and no coding rate is needed to transmit them, dense estimation techniques are made possible, leading to improved results without requiring high coding resources. Experimental results show that the proposed interpolation technique can achieve up to 2.0 dB improvement in SI reconstruction performance, when compared to state-of-the-art DVC techniques.","PeriodicalId":324810,"journal":{"name":"2009 Third ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC)","volume":"23 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132460089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}