{"title":"A low complexity and efficient slice grouping method for H.264/AVC in error prone environments","authors":"Keyu Tan, A. Pearmain","doi":"10.1109/WIAMIS.2009.5031443","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031443","url":null,"abstract":"In this paper, a new method is proposed for Macroblock (MB) importance classification of inter frames. Instead of selecting the most important MBs, the least important MBs are decided first. It makes use of the properties of skip mode in the H.264/AVC standard as the first step. Because the number of MBs chosen as skip mode in a frame varies, further classification is usually required. Four other different features therefore are considered to determine the Important Factor of the remaining MBs. It has been proved that the proposed method can provide good objective and subjective video quality performance, whilst also being simple and fast.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130777956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-class relevance feedback for collaborative image retrieval","authors":"K. Chandramouli, E. Izquierdo","doi":"10.1109/WIAMIS.2009.5031471","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031471","url":null,"abstract":"In recent years, there is an emerging interest to analyse and exploit the log data recorded from different user interactions for minimising the semantic gap problem from multi-user collaborative environments. These systems are referred as “Collaborative Image Retrieval systems”. In this paper, we present an approach for collaborative image retrieval using multi-class relevance feedback. The relationship between users and concepts is derived using Lin Semantic similarity measure from WordNet. Subsequently, the Particle Swarm Optimisation classifier based relevance feedback is used to retrieve similar documents. The experimental results are presented on two well-known datasets namely Corel 700 and Flickr Image dataset. Similarly, the performance of the Particle Swarm Optimised retrieval engine is evaluated against the Genetic Algorithm optimised retrieval engine.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"304 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115378152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Optimizing strategies for the exploration of social networks and associated data collections","authors":"S. Marchand-Maillet, E. Székely, E. Bruno","doi":"10.1109/WIAMIS.2009.5031424","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031424","url":null,"abstract":"Multimedia data collections immersed into social networks may be explored from the point of view of varying documents and users characteristics. In this paper, we develop a unified model to embed documents and users into coherent structures from which to extract optimal subsets. The result is the definition of guiding navigation strategies of both the user and document networks, as a complement to classical search operations. An initial interface that may materialize such browsing over documents is demonstrated in the context of Cultural Heritage.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"31 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121053406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yu Zhou, Melvyn L. Smith, Lyndon N. Smith, R. Warr
{"title":"Segmentation of clinical lesion images using normalized cut","authors":"Yu Zhou, Melvyn L. Smith, Lyndon N. Smith, R. Warr","doi":"10.1109/WIAMIS.2009.5031442","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031442","url":null,"abstract":"Analyzing skin cancer automatically by using image processing techniques has attracted enormous attention recently. The first step in analyzing skin cancer is usually isolating suspicious lesions from normal skin. In this paper, a novel segmentation framework capable of segmenting large clinical lesion images is presented. This algorithm proceeds in a coarse-to-fine approach. Firstly, it builds a down-sampled version of the original image after lower-pass filtering. Then it partitions the down-sampled image by normalized cut. Furthermore, this segmentation result can be adapted to the original image by using a histogram based Bayesian classifier. We also discuss the robustness of this segmentation algorithm with respect to the size of the down-sampled images. Experimental study on synthetic and real images illustrate that this algorithm gives promising results for segmenting clinical lesion images.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126620169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Analysis of affective cues in human-robot interaction: A multi-level approach","authors":"Ginevra Castellano, P. McOwan","doi":"10.1109/WIAMIS.2009.5031482","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031482","url":null,"abstract":"This paper reviews some of the key challenges in affect recognition research for the purpose of designing affect sensitive social robots. An important requirement for a social robot is to be endowed with recognition abilities that vary according to the context of interaction. This paper presents an approach for the analysis of different affective cues depending on the distance at which user and robot interact.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121197961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multicao: A semantic approach to context-aware adaptation decision taking","authors":"V. Barbosa, M. T. Andrade","doi":"10.1109/WIAMIS.2009.5031450","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031450","url":null,"abstract":"The use of context is already a reality in different application fields. One such example is the Robotics field, where sensors are used to capture information about the surrounding environment, allowing the system to react to changes. But also in the multimedia communications field, some work has recently emerged to allow combining sensed context and use that combined information to decide how to adapt the content. This paper describes the MULTICAO ontology, which lies at the core of an Adaptation Decision Engine, providing it with a formal description of the surrounding environment in a similar way as humans would describe it. This way the system has higher chances to take adaptation decisions that best meet users' expectations.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"2 5","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113978867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Toward contextual forensic retrieval for visual surveillance: Challenges and an architectural approach","authors":"Seunghan Han, A. Hutter, W. Stechele","doi":"10.1109/WIAMIS.2009.5031468","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031468","url":null,"abstract":"In traditional visual surveillance systems, retrieval has been relying on indexing events and features extracted by visual analytic algorithms that were developed for well-defined, specific domains. However, due to the increasing need for intelligent forensic retrieval with contextual semantics, this approach is reaching its limits, because it is almost impossible to predict and model all situations at development time. Consequently, a more flexible and intelligent retrieval approach is required. The goal of this paper is to explore the scope of requirements and architectural options to solve this problem. We consider several query scenarios inspired by real events that would benefit from intelligent support. We derive challenges and requirements by reviewing state-of-the-art retrieval approaches in terms of the selected queries. Based on the derived requirements, we present and discuss our architecture and its prototypical implementation.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126396238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Semantic relevance of current image segmentation algorithms","authors":"F. Riaz, M. Dinis-Ribeiro, M. Coimbra","doi":"10.1109/WIAMIS.2009.5031458","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031458","url":null,"abstract":"Several image classification problems are handled using a classical statistical pattern recognition methodology: image segmentation, visual feature extraction, classification. The accuracy of the solution is typically measured by comparing automatic results with manual classification ones, where the distinction between these three steps is not clear at all. In this paper we will focus on one of these steps by addressing the following question: does the visual relevance exploited by segmentation algorithms reflect the semantic relevance of the manual annotation performed by the user? For this purpose we chose a gastroenterology scenario where clinicians classified a set of images into three different types (cancer, pre-cancer, normal), and manually segmented the area they believe was responsible for this classification. Afterwards, we have quantified the performance of two popular segmentation algorithms (mean shift, normalized cuts) on how well they produced one image patch that approximates manual annotation. Results showed that, for this case study, this resemblance is quite close for a large percentage of the images when using normalized cuts.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"130 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134117455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Summarizing raw video material using Hidden Markov Models","authors":"W. Bailer, G. Thallinger","doi":"10.1109/WIAMIS.2009.5031430","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031430","url":null,"abstract":"Besides the reduction of redundancy the selection of representative segments is a core problem when summarizing collections of raw video material. We propose a novel approach for the selection of segments to be included in a video summary based on Hidden Markov Models (HMM), which are trained on an annotated subset of the content. The observations of the HMM are relevance judgments of content segments based on different visual features, the hidden states are the selection/non-selection of content segments. The HMM is designed to take all relevant scenes into account. We show that the approach generalizes well when trained on sufficiently diverse content.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133360612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Summarization of scalable multimedia documents","authors":"Benoît Pellan, C. Concolato","doi":"10.1109/WIAMIS.2009.5031493","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031493","url":null,"abstract":"The summarization of a multimedia document is a challenge that requires the summarization of media elements combined into a document but also relies on an appropriate adaptation of its presentation. In this paper, we present a scalable multimedia model that structures the multimedia scene into incremental Spatial, Temporal and Interactive layers and progressively provides presentation details. Our proposal consists in summarizing such scalable multimedia documents based on three adaptation parameters: a targeted level of expertise, a preferred duration and a level of expectation for extended information. Our approach has been technically validated on PowerPoint-like documents using a generic MPEG-21-based adaptation framework.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"368 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128846630","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}