{"title":"Automatic adaptation decision making using an image to video adaptation tool in the MPEG-21 framework","authors":"Fernando López Hernández, J. Sanchez, N. García","doi":"10.1109/WIAMIS.2009.5031473","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031473","url":null,"abstract":"This paper explains an effective method to combine automatic knowledge-based multimedia adaptation decisions with a quality based decision system that uses regions of interest to improve the outcome of the adaptation. With this purpose, an image to video adaptation tool is used in the experiments.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131720678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"BIlinear Decomposition of 3-D face images: An application to facial expression recognition","authors":"I. Mpiperis, S. Malassiotis, M. Strintzis","doi":"10.1109/WIAMIS.2009.5031417","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031417","url":null,"abstract":"This paper describes a novel technique for decoupling two of the main sources of variation in 3-D facial structure, the subject's identity and expression. Decoupling and controlling independently these factors is a key step in many practical applications and in this work it is achieved by modeling the face manifold with a bilinear model. Bilinear modeling, however, can only be applied to vectors, and therefore a vector representation for each face is established first. To this end, we use a generic face model that is fitted to each face under the constraint that anatomical points get aligned. The effectiveness and applicability of the proposed method is demonstrated with an application to facial expression recognition.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128065467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Self-organizaiton for images from a moving camera","authors":"Yanpeng Cao, J. McDonald","doi":"10.1109/WIAMIS.2009.5031461","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031461","url":null,"abstract":"Given a set of unsorted views captured in a wide area, an effective solution is proposed for image self-organization. The method starts with an initialization step where a small number of key frame pairs are selected to set up a global reference. Given a query image we automatically relate it to the existing key frames based on their pair-wise similarity evaluation. Four major enhancements are made in this step to achieve better performance. Firstly, a recently developed technique, SURF, is applied for robust feature detection. Secondly, an efficient coarse-to-fine matching strategy is implemented. Thirdly, an improved global representation is defined over each image for accurate and fast similarity evaluation. Finally, the method is constantly updated by adding more query images. Experiments were carried out to evaluate the performances of image self-organization by using a large number of images captured from our university's campus.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127459315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Ontological reranking approach for hybrid concept similarity-based video shots indexation","authors":"R. Benmokhtar, B. Huet","doi":"10.1109/WIAMIS.2009.5031474","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031474","url":null,"abstract":"This paper proposes to compare three hybrid concept similarity measures for video shots indexing and retrieval [1], based on two steps. First, individuals concepts are modeled independently. Second, an ontology is introduced via the representation of the relationship between concepts and the ontological readjustment of the confidence values. Our contribution lies in the manner in which inter-concepts similarities are exploited in the indexing system using co-occurrence, visual descriptors, and hybrid semantic similarities. Experimental results report the efficiency and the significant improvement provided by the proposed scheme.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125474890","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Motion-swarm widgets for video interaction","authors":"J. Boyd","doi":"10.1109/WIAMIS.2009.5031441","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031441","url":null,"abstract":"Computer vision systems for human-computer interaction have tended towards more precise forms of interface that require complex vision tasks such as segmentation, tracking, object recognition, pose estimation, and gesture recognition. We present an alternate approach that extrapolates a method for enmasse audience interaction through video. The enmasse interaction simulates a particle moving in the field of motion created by the audience, and the audience interacts by manipulating the particle position. In this paper, we show that by adding sets of constraints to the particle motion, one can build GUI-style widgets. We describe several of these widgets and the results of a small-sample pilot study to test them. The results are not conclusive, but are encouraging, suggesting possibilities for video games and interactive theatre.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116996615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatic topic detection strategy for information retrieval in spoken document","authors":"Shan Jin, Hemant Misra, T. Sikora, J. Jose","doi":"10.1109/WIAMIS.2009.5031492","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031492","url":null,"abstract":"This paper suggests an alternative solution for the task of spoken document retrieval (SDR). The proposed system runs retrieval on multi-level transcriptions (word and phone) produced by word and phone recognizers respectively, and their outputs are combined. We propose to use latent Dirichlet allocation (LDA) model for capturing the semantic information on word transcription. The LDA model is employed for estimating topic distribution in queries and word transcribed spoken documents, and the matching is performed at the topic level. Acoustic matching between query words and phonetically transcribed spoken documents is performed using phone-based matching algorithm. The results of acoustic and topic level matching methods are compared and shown to be complementary.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127728523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A fusion method for multispectral and panchromatic images based on HSI and Contourlet transformation","authors":"Mengxin Song, Xinyu Chen, Ping Guo","doi":"10.1109/WIAMIS.2009.5031436","DOIUrl":"https://doi.org/10.1109/WIAMIS.2009.5031436","url":null,"abstract":"Fusion of multispectral and panchromatic remote sensing images is a procedure to obtain spatial resolution and quality of the panchromatic image as well as preserving spectral information of the multispectral image. In this paper, we present a new fusion method based on HSI (Hue-Saturation-Intensity) and Contourlet transform. First, we convert the multispectral image from the RGB color space into the HSI color space. Then, by applying Contourlet transform to the panchromatic image and the I component of the multispectral image, we utilize an improved fusion rule based on PCA for the low-frequency sub-images, and engage the maximum fusion rule for the high-frequency sub-images. Finally, a fusion image is obtained by the inverse HSI transform. The experimental results show that the proposed fusion method not only enhances the spatial resolution of the fusion image, but also preserves the spectral information of the original multispectral image.","PeriodicalId":233839,"journal":{"name":"2009 10th Workshop on Image Analysis for Multimedia Interactive Services","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128274837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}