S. Georgoulas, G. Pavlou, E. Borcoci, Kin-Hon Ho, E. Vraka
{"title":"Admission Control for End-to-End Multimedia Content Delivery with Quality of Service Guarantees","authors":"S. Georgoulas, G. Pavlou, E. Borcoci, Kin-Hon Ho, E. Vraka","doi":"10.1109/WIAMIS.2008.40","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.40","url":null,"abstract":"End-to-end quality of service guaranteed delivery for multimedia content is a challenging issue, especially in multi-domain environments and heterogeneous network infrastructures. The approach proposed by the ENTHRONE project to solve the end-to-end quality of service problem in a scalable manner is to establish, and activate when needed, long-term quality of service enabled aggregate pipes over multi-domain environments for the subsequent transport of individual flows from multimedia content providers to multimedia content consumers. Based on this approach, this paper proposes admission control schemes both at the granularity of individual flows and aggregate demands. Through simulations we show the proper joint operation of the schemes, their ability to provide quality of service, resource utilization gains and to minimize service rejection probabilities.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132398140","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"On-Line Benchmarking for Multimedia Applications with www.MyMultimediaWorld.com","authors":"Benoît Le Bonhomme, M. Preda, F. Prêteux","doi":"10.1109/WIAMIS.2008.30","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.30","url":null,"abstract":"Here, we present a new method for benchmarking multi-media related algorithms by providing a Web-based architecture able to continuously update (1) the relevant data set through a common representation based on MPEG-4, and (2) the set of algorithms by an open API. The platform benchmark results are updated in real-time and presented for each media, for each group of media or for the entire database. Functionalities for on-line multimedia benchmarking are illustrated for indexation and compression purpose.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116256227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multimedia Indexing and Retrieval of Unedited Audio-Visual Footage","authors":"O. Schreer","doi":"10.1109/WIAMIS.2008.67","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.67","url":null,"abstract":"Unedited audio¿visual footage known as rushes shares many features with general¿purpose multimedia data, but it also shows special characteristics. Rushes are often single¿shot sequences at a single location, sparsely edited, with repetitive content and soundtrack is frequently irrelevant. This leads to additional challenges beyond the existing ones in multimedia indexing and retrieval. The joint effort of a number of research groups all over the world resulted in a ‘rushes exploitation’ task in the TRECVID video analysis international benchmark organized this year. In addition, the European FP6 project RUSHES is fully dedicated to research and development of a system for indexing, accessing and delivering raw, unedited audio¿visual footage and to enable indexing, search and retrieval of rushes archives to ease in¿house postproduction or reuse in a media professional environment. In this special session, latest research results on indexing, search and retrieval with focus on raw unedited audiovisual content are presented.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122872440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Segment Based Diffusion - A Post-Processing Step (Not Only) for Background Subtraction","authors":"M. Unger, M. Asbach","doi":"10.1109/WIAMIS.2008.60","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.60","url":null,"abstract":"A classical approach to video object segmentation is background subtraction. Background subtraction starts by estimating a model of the background image of a scene and then calculating the likeliness that a given pixel of the current camera image belongs to the background model. Typically this is done by subtracting the background image from a given frame, where the difference image is usually thresholded and post-processed to find object boundaries. In this paper we present a method for enhanced post-processing that exploits color and texture information of the original video frame. This way we are able to extract pixel-exact object boundaries. Based on direct color segmentation of the video frame, an iterative method analog to biological diffusion and physical heat transfer processes, spreads information from the difference image over segment boundaries. For this purpose, diffusion resistances are defined between adjacent segments, based on color and texture similarities and common contour length. An iterative process calculates and transfers the flux of 'difference energy' between segments of the difference image. The resulting image allows for easy segmentation by thresholding. Experimental results show the validity of our approach.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126399743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"FCTH: Fuzzy Color and Texture Histogram - A Low Level Feature for Accurate Image Retrieval","authors":"S. Chatzichristofis, Y. Boutalis","doi":"10.1109/WIAMIS.2008.24","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.24","url":null,"abstract":"This paper deals with the extraction of a new low level feature that combines, in one histogram, color and texture information. This feature is named FCTH - Fuzzy Color and Texture Histogram - and results from the combination of 3 fuzzy systems. FCTH size is limited to 72 bytes per image, rendering this descriptor suitable for use in large image databases. The proposed feature is appropriate for accurately retrieving images even in distortion cases such as deformations, noise and smoothing. It is tested on a large number of images selected from proprietary image databases or randomly retrieved from popular search engines. To evaluate the performance of the proposed feature, the averaged normalized modified retrieval rank was used. An online demo that implements the proposed feature in an image retrieval system is available at: http://orpheus.ee.duth.gr/image_retrieval.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124549497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Beilu Shao, D. Renzi, M. Mattavelli, S. Battista, S. Keller
{"title":"A Multimedia Terminal Supporting Adaptation for QoS Control","authors":"Beilu Shao, D. Renzi, M. Mattavelli, S. Battista, S. Keller","doi":"10.1109/WIAMIS.2008.54","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.54","url":null,"abstract":"This paper addresses the end-to-end quality of service (QoS) guarantees and control for multimedia content delivery over heterogeneous networks, with a particular focus on end user terminal perspective. A multimedia terminal which embeds different probes to monitor network conditions and end user perceptual characteristics has been conceived and implemented. A series of QoS mapping between perception and adaptation spaces are yielded by means of the parameters provided by the QoS probes, facilitating content adaptation enabled by any generic adaptation server or intermediate service. Such a design approach illustrates a possible architecture for next-generation multimedia end user system supporting QoS control and content adaptation over a heterogeneous delivery chain.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130464268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Víctor Fernández-Carbajales Cañete, M. García, J. Sanchez
{"title":"Robust People Detection by Fusion of Evidence from Multiple Methods","authors":"Víctor Fernández-Carbajales Cañete, M. García, J. Sanchez","doi":"10.1109/WIAMIS.2008.8","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.8","url":null,"abstract":"This paper describes and evaluates an algorithm for real-time people detection in video sequences based on the fusion of evidence provided by three simple independent people detectors. Experiments with real video sequences show that the proposed integration-based approach is effective, robust and fast by combining simple algorithms.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129651960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Video Content Adaptation Based on SVC and Associated RTP Packet Loss Detection and Signaling","authors":"D. Renzi, P. Amon, S. Battista","doi":"10.1109/WIAMIS.2008.59","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.59","url":null,"abstract":"The development of Scalable Video Coding (SVC) has been directed to serve a wide range of terminals over heterogeneous networks with the same encoded bit stream. SVC provides an elegant solution to adapt the video content, as it allows terminal and/or gateways accessing only a sub-part of the stream, without affecting the semantics of the source video signal. This paper describes adaptation mechanisms based on SVC, focusing on the implementation in the context of the IST project ENTHRONE. A particular emphasis is placed on adaptation techniques based on SVC NAL unit header extension, and on innovative RTP packet loss handling mechanisms, specifically conceived to handle losses in scenarios where SVC to SVC and SVC to H.264/AVC adaptation is performed and layered multicast cannot be used.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129823456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Describing Objects with Multiple Features for Visual Information Retrieval and Annotation","authors":"Qianni Zhang, E. Izquierdo","doi":"10.1109/WIAMIS.2008.45","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.45","url":null,"abstract":"This paper describes how a multi-feature merging approach can be applied in semantic-based visual information retrieval and annotation. The goal is to identify the key visual patterns of specific objects from either static images or video frames. It is shown how the performance of such visual-to-semantic matching schemes can be improved by describing these key visual patterns using particular combinations of multiple visual features. A multi-objective learning mechanism is designed to derive a suitable merging metric for different features. The core of this mechanism is a widely used optimisation method - the multi-objective optimisation strategies. Assessment of the proposed technique has been conducted to validate its performance with natural images and videos.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115538677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Unleashing Video Search","authors":"John R. Smith","doi":"10.1109/WIAMIS.2008.64","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.64","url":null,"abstract":"Video is rapidly becoming a regular part of our digital lives. However, its tremendous growth is increasing userspsila expectations that video will be as easy to search as text. Unfortunately, users are still finding it difficult to find relevant content. And todaypsilas solutions are not keeping pace on problems ranging from video search to content classification to automatic filtering. In this talk we describe recent techniques that leverage the computerpsilas ability to effectively analyze visual features of video and apply statistical machine learning techniques to classify video scenes automatically. We examine related efforts on the modeling of large video semantic spaces and review public evaluations such as TRECVID, which are greatly facilitating research and development on video retrieval. We discuss the role of MPEG-7 as a way to store metadata generated for video in a fully standards-based searchable representation. Overall, we show how these approaches together go a long way to truly unleash video search.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117083732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}