{"title":"Improved side matching for matched-texture coding","authors":"Guoxin Jin, T. Pappas, D. Neuhoff","doi":"10.1109/EUVIP.2014.7018377","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018377","url":null,"abstract":"Matched-texture coding (MTC) exploits the redundancy of textured regions in natural images in order to achieve low-encoding-rate structurally lossless compression. A key element of MTC identifying large image blocks that can be replaced with previously encoded blocks that have similar structure. The side matching (SM) approach attempts to do this by matching the upper and left boundary (side) of a target block with the corresponding boundary of the candidate block, and then, among the best side matches, selecting the one that best matches the target block. We explore the effectiveness of, and interplay between, three SM criteria in order to increase the number and quality of matches and to reduce the computational complexity. The criteria are mean-squared-error, log variance ratio, and partial implementations of STSIM-2, a recently proposed structural texture similarity metric. We propose a hierarchical algorithm for side matching, with three layers that utilize the three metrics, that improves performance and reduces the computation complexity. To set thresholds for the first and second layers of the hierarchical algorithm, we rely on Bayesian hypothesis testing. To estimate the necessary local probability densities, we introduce an adaptive estimation technique that depends on the side matching search region. Experimental results demonstrate an improvement of quality for a given encoding rate over previous realizations of MTC.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129246001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Face recognition with a novel image representation: Local Walsh-Hadamard Transform","authors":"Meryem Uzun-Per, M. Gökmen","doi":"10.1109/EUVIP.2014.7018389","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018389","url":null,"abstract":"In this paper, we present a rich image representation which is robust to illumination, facial expression and scale variations. For this aim, firstly, we propose a novel dense local image representation method based on Walsh Hadamard Transform (WHT) called Local WHT (LWHT). LWHT is the application of WHT to each pixel of an image to decompose it into multiple components, called LWHT maps. Secondly, although LWHT maps are real valued images we propose a method to produce complex valued images from LWHT maps by pairing these maps. We utilize these complex valued image components to obtain Phase Magnitude Histograms (PMHs) in feature extraction stage. Experiments on FERET dataset show that LWHT outperforms Local Binary Patterns (LBP) and Local Gabor Binary Patterns. To further improve the recognition performance, we enhanced our basic method by dividing images into subregions and weighting them, applying cascaded LWHT, and reducing dimension of feature vectors by Block-based Whitened Principal Component Analysis (BWPCA). Experimental results show that the proposed algorithm considerably improves the Walsh-based face recognition and generate comparable results for LBP and Gabor based approaches.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127891067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A new fractional-order variational model for speckled de-noising","authors":"Meriem Hacini, F. Hachouf, K. Djemal","doi":"10.1109/EUVIP.2014.7018384","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018384","url":null,"abstract":"In this paper, a novel speckled image de-noising algorithm is proposed. A fractional-order multiplicative variational model is included as a multiplicative constraint in the regularization problem thereby the appropriate regularization parameter will be controlled by the optimization process itself. An adaptive selection method based on image regions property is used for the selection of the appropriate fractional-order value. The proposed algorithm not only overcomes the disadvantage of generating artificial edges but also has the advantage of de-noising and edges preservation.Experimental results show that the fractional order multiplicative variational model can improve the Peak Signal to Noise Ratio (PSNR) of image, preserve image structures and overcomes the disadvantage of generating artificial edges in the de-noising process.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"528 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124496059","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Vittorio Cuculo, R. Lanzarotti, Giuseppe Boccignone
{"title":"Using sparse coding for landmark localization in facial expressions","authors":"Vittorio Cuculo, R. Lanzarotti, Giuseppe Boccignone","doi":"10.1109/EUVIP.2014.7018369","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018369","url":null,"abstract":"In this article we address the issue of adopting a local sparse coding representation (Histogram of Sparse Codes), in a part-based framework for inferring the locations of facial landmarks. The rationale behind this approach is that unsupervised learning of sparse code dictionaries from face data can be an effective approach to cope with such a challenging problem. Results obtained on the CMU Multi-PIE Face dataset are presented providing support for this approach.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124009447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Weber local descriptor from three orthogonal planes based video copy detection","authors":"Aissa Boukhari, A. Serir","doi":"10.1109/EUVIP.2014.7018374","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018374","url":null,"abstract":"This paper presents a novel video copy detection system. The kernel of the approach is based on our proposed extended local descriptor WLD to three orthogonal planes (WLD-TOP). Indeed, in the aim to extract features vector, key-frames are generated and then a perceptual hash is performed using the WLD-TOP descriptor. The proposed method is applied on three databases and evaluated against several attacks. Performances are assessed in relation to recent studies of the state of the art and the use of other descriptors. The obtained results outline the robustness of the proposed video copy detection in terms of precision, recall and accuracy.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131448299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Segmentation-based 3D dynamic mesh compression scheme","authors":"M. Hachani, A. O. Zaid, W. Puech","doi":"10.1109/EUVIP.2014.7018386","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018386","url":null,"abstract":"This paper presents a novel compression scheme for 3D dynamic meshes. The proposed method mainly uses a segmentation approach which exploits both of temporal and spatial dependencies. The temporal coherence of the geometry component is determined using the heat diffusion properties. The motion of the resulting regions is accurately described by 3D affine transforms. These transforms are computed at the first frame to match the subsequent ones. The temporal prediction errors are then encoded. Comparative coding test, for 3D dynamic meshes, were conducted to evaluate the coding efficiency of the proposed compression scheme. Simulations demonstrate that the rate/distortion results are competitive when compared to the state of the art.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115479248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Chouchane, M. Belahcene, A. Ouamane, S. Bourennane
{"title":"Multimodal face recognition based on histograms of three local descriptors using score level fusion","authors":"A. Chouchane, M. Belahcene, A. Ouamane, S. Bourennane","doi":"10.1109/EUVIP.2014.7018380","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018380","url":null,"abstract":"In this paper, we propose an efficient framework of multimodal face recognition that explores 2D and 3D information based on the score level fusion. To solve the problems of illumination and expression variations, three local methods are introduced, Local Phase Quantization (LPQ), Three-Patch Local Binary Patterns (TPLBP) and Four-Patch Local Binary Patterns (TPLBP). After applying local descriptors to the input image (2D and 3D), this latter is divided into sub-regions or rectangular blocks. Then, the histogram of each sub-region is extracted and concatenated into a single features vector. Principal Component Analysis (PCA) and Enhanced Fisher linear discriminate Model (EFM) are used to reduce the dimensionality. Classification is then performed using the robust Support Vector Machine (SVM) classifier. Finally, score level fusion is used to improve the recognition performance. Experiments are implemented on CASIA3D face database. Our results show that the proposed approach achieves very high performance with RR=98.65% and EER=0.67%.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"10 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131436490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Kinect depth map inpainting using spline approximation","authors":"Denis Brazey, C. Gout","doi":"10.1109/EUVIP.2014.7018372","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018372","url":null,"abstract":"Image inpainting consists in reconstructing missing parts of a given image. In this work, we propose to use an approximation method based on splines and finite elements approximation to recover missing depth values in images acquired with a Kinect 3D sensor. Neighboring pixels in the depth map may contain very different distance values. The considered surface approximation problem therefore involves rapidly varying data which can lead to oscillations (Gibbs phenomenon). To address this issue, we propose to apply two scale transformations to dampen these oscillations near steep gradients implied by the data. The algorithm is presented with some numerical examples of inpainting. Our approach also permits to get a finer resolution of the 3D depth map.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130762443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A new pooling strategy for image quality metrics: Five number summary","authors":"C. Zewdie, Marius Pedersen, Zhaohui Wang","doi":"10.1109/EUVIP.2014.7018373","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018373","url":null,"abstract":"Different pooling strategies for image quality metrics have been proposed over the years. Simple spatial averaging is one of the widely used methods. However, given the fact that a distortion may not be uniformly distributed over the image space, we argue that mean is not a good representation for the image quality distribution. We propose an alternative pooling strategy based on the five number summary. The performance of this pooling strategy has been tested on TID2008, TID2013 and LIVE databases and is shown to produce improvements in correlation with subjective quality assessment.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121097644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D-aided face recognition from videos","authors":"Baptiste Chu, S. Romdhani, Liming Chen","doi":"10.1109/EUVIP.2014.7018366","DOIUrl":"https://doi.org/10.1109/EUVIP.2014.7018366","url":null,"abstract":"The deployment of cameras for security control allows for video stream to be used as input for face recognition (FR). However, most state of the art FR SDKs are generally specifically tuned for dealing with frontal and neutral face images, whereas expression and pose variations, which typically occur in unconstrained settings, e.g., video images, are still major challenges for reliable FR. In this paper, we aim to endow the state of the art FR SDKs with the capabilities to recognize faces in videos. For this purpose, given a video sequence of a person, an extended 3D Morphable Model (3DMM) is used to generate a novel view of this person where the pose is rectified and the expression neutralized. We present a 3DMM fitting method specifically designed for videos to take into account the temporal properties, making use of multiple frames for fitting. Moreover, some constraints of smoothness are used to get a better estimation of its 3D shape and to separate its expression component from its identity component. Finally, we evaluate the proposed method on the Prison Break TV serial and demonstrate its effectiveness using a standard commercial FR SDK.","PeriodicalId":442246,"journal":{"name":"2014 5th European Workshop on Visual Information Processing (EUVIP)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127137676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}