{"title":"Luminance Correction in Stereo Correspondence Based Structure from Motion","authors":"Colin Estermann, W. Stechele, R. Kutka, A. Hutter","doi":"10.1109/WIAMIS.2008.62","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.62","url":null,"abstract":"This paper addresses disparity estimation from image pairs by error minimization. When taken from a monocular camera pan sequence, those images often suffer from luminance variations, caused by the spatially and temporally separated camera positions. Therefore, here we discuss and analyze an existing correction approach for the block matching algorithm. From the observations made we derive a new approach for illumination correction of local error minimization in disparity space. With its simplicity, this multiple disparity space prescaling not only provides the high speed needed for mobile devices, but also decreases the number of false matches by an a priori restriction of correction values.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126464335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Global-to-Local Oriented Rapid Scene Perception","authors":"Le Dong, E. Izquierdo","doi":"10.1109/WIAMIS.2008.12","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.12","url":null,"abstract":"An approach for rapid scene perception from global layout to local features is presented. The representation of a complex scene is initially built from a collection of global features from which properties related to the spatial layout of the scene and its semantic category can be estimated. The rapid perception of natural scenes relies partly on a global estimation of the features contained in the scene. Further analysis on the local essential areas is deployed on the basis. Such kind of integrated model guarantees the interactive processing between local and global features, thus enabling low-level features to initiate scene perception and categorization efficiently.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134431979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Messina, R. Borgotallo, G. Dimino, Daniele Airola Gnota, L. Boch
{"title":"ANTS: A Complete System for Automatic News Programme Annotation Based on Multimodal Analysis","authors":"A. Messina, R. Borgotallo, G. Dimino, Daniele Airola Gnota, L. Boch","doi":"10.1109/WIAMIS.2008.15","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.15","url":null,"abstract":"This paper describes an integrated system for the acquisition, automatic annotation and Web publication of television broadcast news programmes named ANTS (automatic newscast transcription system). The system consists of several analysis components integrated in a unified architecture. Users have the possibility to access a large daily-growing database of news stories from the main national channels, all identified, categorised and published in a fully automatic way. The system identifies story boundaries, extracts texts from spoken content, classifies stories by subject and links external relevant information coming from the Web. Performances of the system have been evaluated in a real-life scenario, by providing an instance of the system to a panel of professional users inside RAI. The strength of the approach behind ANTS consists in the ability to integrate several heterogeneous analysis components in a performant and ready-for-production environment, capable of elaborating many hours of material per day without significant service drops and with sufficiently good accuracy for industrial employment in large broadcasting facilities.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"98 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133418649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
U. Damnjanovic, Virginia Fernandez Arguedas, E. Izquierdo, J. Sanchez
{"title":"Event Detection and Clustering for Surveillance Video Summarization","authors":"U. Damnjanovic, Virginia Fernandez Arguedas, E. Izquierdo, J. Sanchez","doi":"10.1109/WIAMIS.2008.53","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.53","url":null,"abstract":"The target of surveillance summarization is to identify high-value information events in a video stream and to present it to a user. In this paper we present surveillance summarization approach using detection and clustering of important events. Assuming that events are main source of energy change between consecutive frames set of interesting frames is extracted and then clustered. Based on the structure of clusters two types of summaries are created static and dynamic. Static summary is build of key frames that are organized in clusters. Dynamic summary is created from short video segments representing each cluster and is used to lead user to the event of interest captures in key frames. We describe our approach and present experimental results.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133247668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Michael Blighe, N. O’Connor, H. Rehatschek, G. Kienast
{"title":"Identifying Different Settings in a Visual Diary","authors":"Michael Blighe, N. O’Connor, H. Rehatschek, G. Kienast","doi":"10.1109/WIAMIS.2008.17","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.17","url":null,"abstract":"We describe an approach to identifying specific settings in large collections of photographs corresponding to a visual diary. An algorithm developed for setting detection should be capable of clustering images captured at the same real world locations (e.g. in the dining room at home, in front of the computer in the office, in the park, etc.). This requires the selection and implementation of suitable methods to identify visually similar backgrounds in images using their visual features. The goal of the work reported here is to automatically detect settings in images taken over a single week. We achieve this using scale invariant feature transform (SIFT) features and X-means clustering. In addition, we also explore how the use of location based metadata can aid this process.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121405489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xuetao Feng, Yangsheng Wang, Bin Ding, Xiaoyan Wang
{"title":"A Fully Automatic Approach to Facial Feature Tracking Based on Image Registration","authors":"Xuetao Feng, Yangsheng Wang, Bin Ding, Xiaoyan Wang","doi":"10.1109/WIAMIS.2008.6","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.6","url":null,"abstract":"This paper presents a real time, fully automatic facial feature detection and tracking approach. The head pose and facial action is tracked by a modified Candide 3D wireframe model based on an improved image registration technique. An effective model shape and position initialization method is also proposed. Experimental results demonstrate that our system is accurate, robust and fast enough for common applications, even when there are great pose and expression variations.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"71 ","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133488617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Exploiting Spatial Context in Image Region Labelling Using Fuzzy Constraint Reasoning","authors":"C. Saathoff, Steffen Staab","doi":"10.1109/WIAMIS.2008.34","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.34","url":null,"abstract":"We present an approach for integrating explicit knowledge about the spatial context of objects into image region labelling. Our approach is based on spatial prototypes that represent the typical arrangement of objects in images. We use Fuzzy Constraint Satisfaction Problems as the underlying formal model for producing a labelling that is consistent with the spatial constraints of prototypes.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133960401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Foreground Detection Using the Choquet Integral","authors":"Fida El Baf, T. Bouwmans, B. Vachon","doi":"10.1109/WIAMIS.2008.9","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.9","url":null,"abstract":"Foreground Detection is a key step in background subtraction problem. This approach consists in the detection of moving objects from static cameras through a classification process of pixels as foreground or background. The presence of some critical situations i.e noise, illumination changes and structural background changes produces an uncertainty in the classification of image pixels which can generate false detections. In this context, we propose a fuzzy approach using the Choquet integral to avoid the uncertainty in the classification. The experiments on different video datasets have been realized by testing different color space and by fusing color and texture features. The proposed method is characterized through robustness against illumination changes, shadows and little background changes, and it is validated with the experimental results.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116550020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Examining Kalman Filters Applied to Tracking Objects in Motion","authors":"A. Czyżewski, P. Dalka","doi":"10.1109/WIAMIS.2008.23","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.23","url":null,"abstract":"Kalman filters were used for establishing relations between objects moving in video frames to the real moving objects under analysis. As a result of applying some popular methods of moving objects detection, the objects were represented by rectangles. A two-dimensional colour histogram based on a chromatic space was used for each object in experiments. The objects coupling with adequate regions including the relation of many-to-many was studied experimentally employing Kalman filters. The implemented algorithm provides a part of an advanced audio-video surveillance system for security applications.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115924151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatically Segmenting LifeLog Data into Events","authors":"A. Doherty, A. Smeaton","doi":"10.1109/WIAMIS.2008.32","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.32","url":null,"abstract":"A personal lifelog of visual information can be very helpful as a human memory aid. The SenseCam, a passively capturing wearable camera, captures an average of 1785 images per day, which equates to over 600000 images per year. So as not to overwhelm users it is necessary to deconstruct this substantial collection of images into digestable chunks of information, i.e. into distinct events or activities. This paper improves on previous work on automatic segmentation of SenseCam images into events by up to 29.2%, primarily through the introduction of intelligent threshold selection techniques, but also through improvements in the selection of normalisation, fusion, and vector distance techniques. Here we use the most extensive dataset ever used in this domain, 271163 images collected by 5 users over a time period of one month with manually groundtruthed events.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114910286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}