{"title":"On Design of Linear Minimum-Entropy Predictor","authors":"X. Wang, Xiaolin Wu","doi":"10.1109/MMSP.2007.4412852","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412852","url":null,"abstract":"Linear predictors for lossless data compression should ideally minimize the entropy of prediction errors. But in current practice predictors of least-square type are used instead. In this paper, we formulate and solve the linear minimum-entropy predictor design problem as one of convex or quasiconvex programming. The proposed minimum-entropy design algorithms are derived from the well-known fact that prediction errors of most signals obey generalized Gaussian distribution. Empirical results and analysis are presented to demonstrate the superior performance of the linear minimum-entropy predictor over the traditional least-square counterpart for lossless coding.","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121734825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Wavelet-Based Multi-View Video Coding with Spatial Scalability","authors":"Jens-Uwe Garbas, A. Kaup","doi":"10.1109/MMSP.2007.4412906","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412906","url":null,"abstract":"In this paper, we propose two wavelet-based frameworks which allow fully scalable multi-view video coding. Using a 4-D wavelet transform, both schemes generate a bitstream that can be truncated to achieve a temporally, view-directionally, and/or spatially downscaled representation of the coded multi-view video sequence. Well-known wavelet-based scalable coding schemes for single-view video sequences have been adopted and extended to match the specific needs of scalable multi-view video coding. Motion compensated temporal filtering (MCTF) is applied to each Video sequence of each camera to exploit temporal correlation and inter-view dependencies are exploited with disparity compensated view filtering (DCVF). A spatial wavelet transform is utilized either before and after temporal-view-directional decomposition (2D+T+V+2D scheme) or only after the temporal-view-directional decomposition (T+V+2D scheme) for spatial decorrelation. The influence of the two different approaches on spatial scalability is shown in this paper as well as the superior coding efficiency of both codecs compared with simulcast coding.","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126259513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Cross-Layer Adaptive ARQ for Uplink Video Streaming in Tandem Wireless/Wireline Networks","authors":"A. Argyriou","doi":"10.1109/MMSP.2007.4412840","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412840","url":null,"abstract":"In this paper, we focus on improving the robustness of packetized multimedia streaming in tandem-connected wireless LANs and wireline packet switched networks. To this aim we initially develop an analytical model that expresses the end-to-end packet loss rate and latency, as a function of the retransmission-based error control mechanisms employed both at the application and wireless link layers. The developed model is the basis of an algorithm that dynamically identifies the optimal number of retransmissions at each protocol layer, so that the overall effective packet loss rate is minimized. Realistic video streaming experiments show considerable quality improvements in terms of PSNR, by avoiding the overall number of retransmissions.","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"101 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131830410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Symmetric Distributed Arithmetic Coding of Correlated Sources","authors":"Marco Grangetto, E. Magli, G. Olmo","doi":"10.1109/MMSP.2007.4412830","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412830","url":null,"abstract":"We propose a new scheme for symmetric Slepian-Wolf coding of correlated binary sources. Unlike previous designs that employ capacity-achieving channel codes, the proposed scheme is based on arithmetic codes with error correction capability. We define a time-sharing version of a distributed arithmetic coder, and a soft joint decoder. Experimental results on two sources show that, for short block length, the proposed scheme outperforms the symmetric turbo code design in (Stankovic et al., 2006).","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134640961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multiscale Integral Invariants For Facial Landmark Detection in 2.5D Data","authors":"Adam Slater, Y. Hu, N. Boston","doi":"10.1109/MMSP.2007.4412846","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412846","url":null,"abstract":"In this paper, we introduce a novel 3D surface landmark detection method using a 3D integral invariant feature extended from that proposed by Manay et al. for 2D contours. We apply this new feature to detect the nose tips of 2.5D range images of human faces. Using the Face Recognition Grand Challenge 2.0 dataset, our method compares favorably with a recently proposed competing method.","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134141892","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Content-based Video Signatures based on Projections of Difference Images","authors":"R. Radhakrishnan, C. Bauer","doi":"10.1109/MMSP.2007.4412886","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412886","url":null,"abstract":"We propose a novel video signature extraction method based on projections of difference images between consecutive video frames. The difference images are projected onto random basis vectors to create a low dimensional bitstream representation of the active content (moving regions) between two video frames. A sequence of these signatures serves to identify the underlying video content in a robust manner. Our experimental results show that the proposed video signature is robust to most common signal processing operations on video content such as compression, resolution scaling, brightness scaling.","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133278352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Analysis of multimodal binary detection systems based on dependent/independent modalities","authors":"O. Koval, S. Voloshynovskiy, T. Pun","doi":"10.1109/MMSP.2007.4412820","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412820","url":null,"abstract":"Performance limits of multimodal detection systems are analyzed in this paper. Two main setups are considered, i.e., based on fusion of dependent and independent modalities, respectively. The analysis is performed in terms of attainable probability of detection errors characterized by the corresponding error exponents. It is demonstrated that an expected performance gain from fusion of dependent modalities is superior than in the case when one fuses independent signals. In order to quantify the efficiency of dependent modality fusion versus the independent case, the problem analysis is performed in the Gaussian formulation.","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134083589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Combining Vocal Source and MFCC Features for Enhanced Speaker Recognition Performance Using GMMs","authors":"Danoush Hosseinzadeh, S. Krishnan","doi":"10.1109/MMSP.2007.4412892","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412892","url":null,"abstract":"This work presents seven novel spectral features for speaker recognition. These features are the spectral centroid (SC), spectral bandwidth (SBW), spectral band energy (SBE), spectral crest factor (SCF), spectral flatness measure (SFM), Shannon entropy (SE) and Renyi entropy (RE). The proposed spectral features can quantify some of the characteristics of the vocal source or the excitation component of speech. This is useful for speaker recognition since vocal source information is known to be complementary to the vocal tract transfer function, which is usually obtained using the Mel frequency cepstral coefficients (MFCC) or linear predication cepstral coefficients (LPCC). To evaluate the performance of the spectral features, experiments were performed using a text-independent cohort Gaussian mixture model (GMM) speaker identification system. Based on 623 users from the TIMIT database, the spectral features achieved an identification accuracy of 99.33% when combined with the MFCC based features and when using undistorted speech. This represents a 4.03% improvement over the baseline system trained with only MFCC and DeltaMFCC features.","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132725997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Experiments in Automatic Genre Classification of Full-length Music Tracks using Audio Activity Rate","authors":"Shiva Sundaram, Shrikanth S. Narayanan","doi":"10.1109/MMSP.2007.4412827","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412827","url":null,"abstract":"The activity rate of an audio clip in terms of three defined attributes results in a generic, quantitative measure of various acoustic sources present in it. The objective of this work is to verify if the acoustic structure measured in terms of these three attributes can be used for genre classification of music tracks. For this, we experiment on classification of full-length music tracks by using a dynamic time warping approach for time-series similarity (derived from the activity rate measure) and also a Hidden Markov Model based classifier. The performance of directly using timbral (Mel-frequency Cepstral Coefficients) features is also presented. Using only the activity rate measure we obtain classification performance that is about 35% better than baseline chance and this compares well with other proposed systems that use musical information such as beat histogram or pitch based melody information.","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125137805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Makrogiannis, J. Wellen, Y. Wu, L. Bloy, S. Sarkar
{"title":"A Multimodal Image Registration and Fusion Methodology Applied to Drug Discovery Research","authors":"S. Makrogiannis, J. Wellen, Y. Wu, L. Bloy, S. Sarkar","doi":"10.1109/MMSP.2007.4412883","DOIUrl":"https://doi.org/10.1109/MMSP.2007.4412883","url":null,"abstract":"The development of novel methodologies that utilize various non-invasive imaging modalities has resulted in their increased use and relevance in the pie-clinical phase of pharmaceutical compound development. Scientific questions that may benefit from the use of imaging techniques are often more robustly addressed with data acquired from complementary imaging modalities, intensifying the need for performing cross-modality image registration and fusion. In this work, a methodology for multimodal coregistration and fusion of MRI and PET volumes is presented with the aim of visualizing the biodistribution of a radiolabeled compound (from PET) in a high-resolution anatomical reference image (from MRI). The use of an animal platform and fiducial markers for defining spatial correspondences is considered first, followed by a preliminary automated alignment approach that calculates and matches shape-based features using Parzen estimators and genetic algorithm optimization. Experimental results are also presented, followed by concluding remarks and future perspectives on our work.","PeriodicalId":225295,"journal":{"name":"2007 IEEE 9th Workshop on Multimedia Signal Processing","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125540611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}