Michael Eberhard, Luca Celetto, C. Timmerer, E. Quacchio, H. Hellwagner
{"title":"Performance Analysis of Scalable Video Adaptation: Generic versus Specific Approach","authors":"Michael Eberhard, Luca Celetto, C. Timmerer, E. Quacchio, H. Hellwagner","doi":"10.1109/WIAMIS.2008.48","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.48","url":null,"abstract":"This paper provides a performance analysis of adaptation approaches designed for scalable media resources. In particular, we investigate the streaming of media resources compliant to the scalable video coding (SVC) extensions of advanced video coding (AVC) within heterogeneous environments, i.e., terminals and networks with different capabilities. Therefore, we have developed a test-bed in order to analyze two different approaches for the adaptation of scalable media resources, namely a generic approach that is applicable independently of the actual scalable coding format used and a specific approach especially built for SVC. The results show that if adaptation is required the generic approach clearly outperforms the approach specifically built for SVC.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123058090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D Inference and Modelling for Video Retrieval","authors":"Huiyu Zhou, A. Sadka, Richard M. Jiang","doi":"10.1109/WIAMIS.2008.37","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.37","url":null,"abstract":"A new scheme is proposed for extracting planar surfaces from 2D image sequences. We firstly perform feature correspondence over two neighboring frames, followed by the estimation of disparity and depth maps, provided a calibrated camera. We then apply iterative random sample consensus (RANSAC) plane fitting to the generated 3D points to find a dominant plane in a maximum likelihood estimation style. Object points on or off this dominant plane are determined by measuring their Euclidean distance to the plane. Experimental work shows that the proposed scheme leads to better plane fitting results than the classical RANSAC method.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125015935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Revisiting the Vector Retrieval Model in Context of the MPEG-7 Semantic Description Scheme","authors":"M. Lux","doi":"10.1109/WIAMIS.2008.19","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.19","url":null,"abstract":"Due to the semantic gap multimedia metadata is currently the only way to allow multimedia information retrieval on a conceptual level. One prominent multimedia metadata standard is MPEG-7. It offers means to annotate multimedia content using concepts and semantic relations. While several practical solutions for retrieval of MPEG-7 based conceptual annotations exist, a basic theoretical model has not been discussed yet. The paper in hand presents a vector representation for the MPEG-7 semantic description scheme and shows that the vector representation is unique and invertible. The vector representation allows the application of the vector retrieval model and therefore common retrieval mechanisms like weighting and indexing techniques or clustering.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122308855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"On-Line Video Summarization Based on Signature-Based Junk and Redundancy Filtering","authors":"Víctor Valdés, J. Sanchez","doi":"10.1109/WIAMIS.2008.7","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.7","url":null,"abstract":"The huge amount of video content and the increasing rate in which new content is being created makes video summarization a necessary tool for the efficient access and management of video repositories. Nevertheless, the huge amount of data and the complexity of current techniques make the implantation of the summarization systems a difficult or non practical task. This paper presents current advances in the development of a novel algorithm for on-line video summarization, which can be easily used also as a video retrieval tool, aimed to combine summarization quality with high computational efficiency providing a system that can be practical for both commercial and home environments.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117008018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ville Viitaniemi, Mats Sjöberg, M. Koskela, Jorma T. Laaksonen
{"title":"Exploiting Temporal and Inter-concept Co-occurrence Structure to Detect High-Level Features in Broadcast Videos","authors":"Ville Viitaniemi, Mats Sjöberg, M. Koskela, Jorma T. Laaksonen","doi":"10.1109/WIAMIS.2008.50","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.50","url":null,"abstract":"In this paper the problem of detecting high-level features from video shots is studied. In particular, we explore the possibility of taking advantage of temporal and interconcept co-occurrence patterns that the high-level features of a video sequence exhibit. Here we present two straightforward techniques for the task: N-gram models and clustering of temporal neighbourhoods. We demonstrate the usefulness of these techniques on data sets of the TRECVID high-level feature detection tasks of the years 2005-2007.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"174 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132561310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"MobSOS - A Testbed for Mobile Multimedia Community Services","authors":"D. Renzel, R. Klamma, M. Spaniol","doi":"10.1109/WIAMIS.2008.52","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.52","url":null,"abstract":"Due to recent developments in the domain of mobile broadband communication and a growing availability of standardized development tools for mobile devices a new generation of mobile services and applications is expected to emerge in the near future combining well-established multimedia and community concepts with mobile aspects. However, it is often challenging to predict the success of such new services. In this paper we present a short overview of the MobSOS testbed and its homonymous success model based on a combination of traditional information system success models with modern requirements for mobile multimedia communities. Finally we outline the application of the MobSOS testbed to NMV Mobile, a mobile multimedia capturing, annotation and retrieval tool.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131046811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Real-Time People Counting Using Multiple Lines","authors":"Javier Barandiarán, Berta Murguia, F. Boto","doi":"10.1109/WIAMIS.2008.27","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.27","url":null,"abstract":"A novel real-time people counting system is presented in this paper. Using a single overhead mounted camera, the system counts the number of people going in and out of an observed area. Counting is performed by analyzing an image zone composed by a set of virtual counting lines. The system runs on a commercial PC, does not need a special background and is easily adjustable to different camera height requirements. We have tested the performance of the system, achieving a correct people counting rate of 95%.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122112559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Roman Kern, M. Granitzer, Viktoria Pammer-Schindler
{"title":"Extending Folksonomies for Image Tagging","authors":"Roman Kern, M. Granitzer, Viktoria Pammer-Schindler","doi":"10.1109/WIAMIS.2008.43","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.43","url":null,"abstract":"Due to the unsatisfactory results of content based image retrieval methods, organisation and retrieval of multimedia data strongly relies on metadata and free text description. Folksonomies, collaboratively created sets of metadata, merged recently and help organising multimedia information on the Web. Our contribution addresses the question how to extend a classical folksonomy with additional metadata. We also discuss the quality of the extended folksonomy and its application for tag recommendation. We show that some relations of the original folksonomies can be replaced while others are unique. In addition our analysis shows, that for 40% the correct tag is in the first 10% of the set of tag recommendations.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129386546","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
O. Schreer, Leticia Fuentes Ardeo, Dimitrious Sotiriou, A. Sadka, E. Izquierdo
{"title":"User Requirements for Multimedia Indexing and Retrieval of Unedited Audio-Visual Footage - RUSHES","authors":"O. Schreer, Leticia Fuentes Ardeo, Dimitrious Sotiriou, A. Sadka, E. Izquierdo","doi":"10.1109/WIAMIS.2008.14","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.14","url":null,"abstract":"Multimedia analysis and reuse of raw un-edited audio visual content known as rushes is gaining acceptance by a large number of research labs and companies. A set of research projects are considering multimedia indexing, annotation, search and retrieval in the context of European funded research, but only the FP6 project RUSHES is focusing on automatic semantic annotation, indexing and retrieval of raw and un-edited audio-visual content. Even professional content creators and providers as well as home-users are dealing with this type of content and therefore novel technologies for semantic search and retrieval are required. As a first result of this project, the user requirements and possible user-scenarios are presented in this paper. These results lay down the foundation for the research and development of a multimedia search engine particularly dedicated to the specific needs of the users and the content.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128696131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Measuring of Flickering Artifacts in Predictive Coded Video Sequences","authors":"J. Pandel","doi":"10.1109/WIAMIS.2008.47","DOIUrl":"https://doi.org/10.1109/WIAMIS.2008.47","url":null,"abstract":"Flickering is one of the most annoying temporal artifacts in predictive video coding. It will frequently appear in particular within slow moving low textured objects when the encoder is forced to skip macroblocks in order to save data rate. A very simple no-reference metric is derived for automatic detection of temporal flickering artifacts resulting from such macroblock skipping. Tests are performed for different H.264 encoded video sequences.","PeriodicalId":325635,"journal":{"name":"2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130015063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}