Vasilios Anagnostopoulos, N. Doulamis, A. Doulamis
{"title":"Edge-motion video summarization: Economical video summarization for low powered devices","authors":"Vasilios Anagnostopoulos, N. Doulamis, A. Doulamis","doi":"10.1109/WIAMIS.2009.5031488","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031488","url":null,"abstract":"In this paper we propose a new dynamic video summarization procedure. For this purpose, we find an economical image representation in terms of edge-like features that facilitates video re-sampling with minimal loss of information. We propose a fast noise tolerant way to extract the representation. We argue against the use of an auxiliary descriptor for the samples and we show how we can re-use the existing Principal Components Analysis (PCA) with the appropriate modifications on the new image representation without sacrificing accuracy.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126967968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-resolution tracking of a non-rigid target with particle filters for low and variable frame-rate videos","authors":"Yifan Zhou, J. Benois-Pineau, H. Nicolas","doi":"10.1109/WIAMIS.2009.5031444","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031444","url":null,"abstract":"For the last decades, there has been a focus on using particle filters methods in tracking systems for their ability to solve nonlinear, non-Gaussian tracking problems. This article presents a novel tracking method with multi-resolution particle filters using color appearance model for tracking of a non-rigid target such as people. Compared to the baseline tracking by particle filters, our approach has the advantage of a two passes tracking: fast locating target state in the 1st pass and precisely refining it in the 2nd pass. After being evaluated in different kinds of people tracking environments, our system proves to offer a robust, efficient tracking performance and less computational cost.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131038577","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The fast-3D spatio-temporal interest region detector","authors":"Sander Koelstra, I. Patras","doi":"10.1109/WIAMIS.2009.5031478","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031478","url":null,"abstract":"Spatio-temporal interest region detectors can be used in the analysis of video to determine sparse, informative regions as candidates for feature extraction. In this paper we compare existing detectors and introduce the new FAST-3D detector, loosely based on the FAST spatial interest region detector. We compare the invariance of detectors to rotation, scale and compression by measuring the similarity between detected interest regions in original and transformed versions of videos. We measure both the repeatibility and introduce a new similarity measure based on mutual information. The FAST-3D detector is shown to be on par with the other detectors, while showing a significant increase in speed.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125325471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Face tracking using a region-based mean-shift algorithm with adaptive object and background models","authors":"Verónica Vilaplana, David Varas","doi":"10.1109/WIAMIS.2009.5031419","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031419","url":null,"abstract":"This paper proposes a technique for face tracking based on the mean shift algorithm and the segmentation of the images into regions homogeneous in color. Object and background are explicitly modeled and updated through the tracking process. Color and shape information are used to define with precision the face contours, providing a mechanism to adapt the tracker to variations in object scale and to illumination and background changes.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"116 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123164425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Autonomous production of basketball videos from multi-sensored data with personalized viewpoints","authors":"Fan Chen, C. Vleeschouwer","doi":"10.1109/WIAMIS.2009.5031437","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031437","url":null,"abstract":"We propose an autonomous system for personalized production of basketball videos from multi-sensored data under limited display resolution. We propose criteria for optimal planning of viewpoint coverage and camera selection for improved story-telling and perceptual comfort. By using statistical inference, we design and implement the estimation process. Experiments are made to verify the system, which shows that our method efficiently alleviates flickering visual artifacts due to viewpoint switching, and discontinuous story-telling artifacts.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122680703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Context awareness in graph-based image semantic segmentation via visual word distributions","authors":"G. Passino, I. Patras, E. Izquierdo","doi":"10.1109/WIAMIS.2009.5031425","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031425","url":null,"abstract":"This paper addresses the problem of image semantic segmentation (or semantic labelling), that is the association of one of a predefined set of semantic categories (e.g. cow, car, face) to each image pixel. We adopt a patch-based approach, in which super-pixel elements are obtained via oversegmentation of the original image. We then train a Conditional Random Field on heterogeneous descriptors extracted at different scales and locations. This discriminative graphical model can effectively account for the statistical dependence of neighbouring patches. For the more challenging task of considering long-range patch dependency and contextualisation, we propose the use of a descriptor based on histograms of visual words extracted in the vicinity of each patch at different scales. Experiments validate our approach by showing improvements with respect to both a base model not using distributed features and the state of the art works in the area.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132112605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Evaluation of quality scalability layer selection for bit rate adaptation of scalable video content","authors":"G. Nur, H. K. Arachchi, S. Dogan, A. Kondoz","doi":"10.1109/WIAMIS.2009.5031472","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031472","url":null,"abstract":"Scalable Video Coding (SVC), which offers a comprehensive set of scalability options including spatial, temporal, and quality scalabilities, is considered to be one of the core enabling technologies for realising the Universal Multimedia Access (UMA). Efficiently utilising the functionalities of these options to satisfy the demand of better visual quality for collaborative networking environments is an important issue that needs attention. In this paper, we investigate the effects of asymmetrically adjusting the balance of quality enhancement layers in key and non-key frames on Rate-Distortion (R-D) performance of the adapted video bit streams. Experimental results presented in the paper demonstrate that the asymmetric bit rate adaptation scheme is capable of not only improving the R-D performance but also increasing the number of achievable bit rate points.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131420281","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Archive film defect detection based on a hidden Markov model","authors":"Xiaosong Wang, M. Mirmehdi","doi":"10.1109/WIAMIS.2009.5031490","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031490","url":null,"abstract":"We propose a novel statistical approach to detect defects in digitized archive film by using temporal information across a number of frames modeled with an HMM. The HMM is trained for normal observation sequences and then applied within a framework to detect defective pixels by examining each new observation sequence and its subformations via a leave-one-out process. We compare against state-of-the-art results to demonstrate that the proposed method achieves better detection rates, with fewer false alarms.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131466264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Anisotropic fluid solver for robust optical flow smoothing","authors":"A. Doshi, A. Bors","doi":"10.1109/WIAMIS.2009.5031446","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031446","url":null,"abstract":"Computational fluid dynamics provides the framework for explaining fluid motion. The methodology proposed in this paper applies partial differential equations such as Navier-Stokes for modelling the optical flow in image sequences displaying complex motion. The proposed robust stable fluid solver has the following components: robust diffusion, advection and mass conservation. We employ a robust diffusion kernel which combines the geometry preserving property of the heat kernel with an outlier rejection mechanism. The proposed methodology is applied on the artificially generated Von Karman flows, after considering additive noise, and onto the optical flow extracted from real image sequences.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116503019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Effect of H.264/AVC compression on object detection for video surveillance","authors":"C. Poppe, S. D. Bruyne, P. Lambert, R. Walle","doi":"10.1109/WIAMIS.2009.5031449","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031449","url":null,"abstract":"Nowadays, video surveillance systems apply video compression to reduce bandwith and storage cost. However, generally, these video sequences are the input, after decoding, for video analysis modules. H.264/AVC is the newest video standard and is assumed to be omnipresent in video surveillance systems in the near future. Since the video compression introduces artefacts, which influence the performance of these analysis modules, it is important to make a quantitative evaluation of this effect. Hence, in this paper we present the first quantitative analysis of the effect that H.264/AVC compression has upon a generally accepted moving object detection technique. We analyze different encoding schemes and show the influence on the object detection results for different representative sequences.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114528411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}