{"title":"Cascade scheme face detection using a non-liniar classifier","authors":"A. Rama, F. Tarrés, A. Soria-Frisch","doi":"10.1109/WIAMIS.2009.5031455","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031455","url":null,"abstract":"In this paper, the non-linear fuzzy integral operator is proposed for combining different sets of Haar features for face detection. The proposed method presents a lower false detection rate than the State-of-the-art AdaBoost face detector by a similar true acceptance rate and using the same optimal set of features. Furthermore, this novel face detector seems to have a better generalization capability than the AdaBoost method. Experimental results show a positive face detection rate larger than 92% and a false detection rate of 0.1% when using a four stage cascade scheme.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122430859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Detection of pan and zoom in soccer sequences based on H.264/AVC motion information","authors":"L. Superiori, M. Rupp","doi":"10.1109/WIAMIS.2009.5031427","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031427","url":null,"abstract":"Unsupervised detection of pan and zoom in soccer sequences allows automatic classification of shots and match analysis. In this work we propose a pan and zoom (both in and out) detector specifically designed for low resolution soccer sequences. Our implementation is based on the analysis of the distribution of the motion vectors, already available in the encoded sequence, among a specific subset of reliable MBs, selected by means of inexpensive image preprocessing.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115899267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Building summaries from web information sources","authors":"A. d’Acierno, V. Moscato, A. Picariello","doi":"10.1109/WIAMIS.2009.5031431","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031431","url":null,"abstract":"Document summarization techniques can be profitably used for automatic production and delivery of multimedia information. In this paper we describe a system for summarizing HTML documents (retrieved from the Internet) using several heuristic optimization criteria. An overview of the system and some preliminary results are described.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131413556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
E. Rossi, Sergio Benini, R. Leonardi, Boris Mansencal, J. Benois-Pineau
{"title":"Clustering of scene repeats for essential rushes preview","authors":"E. Rossi, Sergio Benini, R. Leonardi, Boris Mansencal, J. Benois-Pineau","doi":"10.1109/WIAMIS.2009.5031476","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031476","url":null,"abstract":"This paper focuses on a specific type of unedited video content, called rushes, which are used for movie editing and usually present a high-level of redundancy. Our goal is to automatically extract a summarized preview, where redundant material is diminished without discarding any important event. To achieve this, rushes content has been first analysed and modeled. Then different clustering techniques on shot key-frames are presented and compared in order to choose the best representative segments to enter the preview. Experiments performed on TRECVID data are evaluated by computing the mutual information between the obtained results and a manually annotated ground-truth.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114399938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Poslad, Aristodemos Pnevmatikakis, M. Nunes, Elena Garrido Ostermann, P. Chippendale, P. Brightwell, C. Patrikakis
{"title":"Directing your own live and interactive sports channel","authors":"S. Poslad, Aristodemos Pnevmatikakis, M. Nunes, Elena Garrido Ostermann, P. Chippendale, P. Brightwell, C. Patrikakis","doi":"10.1109/WIAMIS.2009.5031486","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031486","url":null,"abstract":"The ability to mark-up live sports event content, viewed from multiple camera angles, such that athletes and other objects of interest can be tracked, facilitates an exciting new personalised and interactive viewing experience for spectators, enabling spectators to act as directors of their own customised live sports videos. In this paper, such an approach is described as part of the My-e-Director 2012 project. The design of this platform is described here and a discussion of a prototype system is given.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116833280","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
G. Caridakis, K. Karpouzis, Athanasios I. Drosopoulos, S. Kollias
{"title":"Adaptive gesture recognition in Human Computer Interaction","authors":"G. Caridakis, K. Karpouzis, Athanasios I. Drosopoulos, S. Kollias","doi":"10.1109/WIAMIS.2009.5031485","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031485","url":null,"abstract":"An adaptive, invariant to user performance fluctuation or noisy input signal, gesture recognition scheme is presented based on Self Organizing Maps, Markov Models and Levenshtein sequence distance. Multiple modalities, all based on the hand position during gesturing, train different classifiers which are then fused in a weak classifier boosting-like setup by weight assignment to each stream. The adaptability of the proposed approach consists of the incorporation of Self Organizing Maps during training, the exploitation of neighboring relations between states of the Markov models and the modified Levenshtein distance algorithm. The main focus of current work is to tackle intra and inter user variability during gesture performance by adding flexibility to the decoding procedure and allowing the algorithm to perform an optimal trajectory search while the processing speed of both the feature extraction and the recognition process indicate that the proposed architecture is appropriate for real time and large scale lexicon applications.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"219 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123029228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Generic colour image segmentation via multi-stage region merging","authors":"Gaurav Gupta, A. Psarrou, A. Angelopoulou","doi":"10.1109/WIAMIS.2009.5031464","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031464","url":null,"abstract":"We present a non-parametric unsupervised colour image segmentation system that is fast and retains significant perceptual correspondence with the input data. The method uses a region merging approach based on statistics of growing local structures. A two-stage algorithm is employed during which neighbouring regions of homogeneity are traced using feature gradients between groups of pixels, thus giving priority to topological relations. The system finds spatially cohesive and globally salient image regions usually without losing smaller localised areas of high saliency. Unoptimised implementations of the method work nearly in real-time, handling multiple frames a second. The system is successfully applied to problems such as object detection and tracking.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124758311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Evaluation of pixel- and motion vector-based global motion estimation for camera motion characterization","authors":"M. Haller, A. Krutz, T. Sikora","doi":"10.1109/WIAMIS.2009.5031429","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031429","url":null,"abstract":"Pixel-based and motion vector-based global motion estimation (GME) techniques are evaluated in this paper with an automatic system for camera motion characterization. First, the GME techniques are compared with a frame-by-frame PNSR measurement using five video sequences. The best motion vector-based GME method is then evaluated together with a common and a simplified pixel-based GME technique for camera motion characterization. For this, selected unedited videos from the TRECVid 2005 BBC rushes corpus are used. We evaluate how the estimation accuracy of global motion parameters affects the results for camera motion characterization in terms of retrieval measures. The results for this characterization show that the simplified pixel-based GME technique obtains results that are comparable with the common pixel-based GME method, and outperforms significantly the results of an earlier proposed motion vector-based GME approach.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124799111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Event detection in athletics for personalized sports content delivery","authors":"N. Katsarakis, Aristodemos Pnevmatikakis","doi":"10.1109/WIAMIS.2009.5031487","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031487","url":null,"abstract":"Broadcasting of athletics is nowadays biased towards running (sprint and longer distances) sports. Personalized content delivery can change that for users that wish to focus on different content. Using a combination of video signal processing algorithms and live information that accompanies the video of large-scale sports like the Olympics, a system can attend to the preferences of users by selecting the most suitable camera view for them.There are two types of camera selection for personalized content delivery. According to the between sport camera selection, the view is changed between two sports, upon the onset of a sport higher up the user preferences than the one currently being delivered. According to the within sport camera selection, the camera is changed to offer a better view of the evolution of the sport, based on the phase it is in. This paper details the video processing algorithms needed for the extraction of the events that trigger both between and within sport camera selection, and describes a system that handles user preferences, live information and video-generated events to offer personalized content to the users.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125577615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Exploiting visual reranking to improve pseudo-relevance feedback for spoken-content-based video retrieval","authors":"S. Rudinac, M. Larson, A. Hanjalic","doi":"10.1109/WIAMIS.2009.5031421","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031421","url":null,"abstract":"In this paper we propose an approach that utilizes visual features and conventional text-based pseudo-relevance feedback (PRF) to improve the results of semantic-theme-based video retrieval. Our visual reranking method is based on an Average Item Distance (AID) score. AID-based visual reranking is designed to improve the suitability of items at the top of the initial results list, i.e., those feedback items selected for use in query expansion. Our method is intended to help target feedback items representative of visual regularity typifying the semantic theme of the query. Experiments performed on the VideoCLEF 2008 data set and on a number of retrieval scenarios combining the inputs from speech-transcript-based (i.e., text-based) search and visual reranking demonstrate the benefits of using AID-based visual representatives to compensate for the inherent problems of PRF, such as topic drift.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128729765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}