{"title":"Compact representation of images by edge adapted multiscale transforms","authors":"A. Cohen, Basarab Matei","doi":"10.1109/ICIP.2001.958938","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958938","url":null,"abstract":"We introduce new multiscale representations for images which incorporate a specific geometric treatment of edges. The associated transforms are inherently nonlinear and nontensor product, in contrast to classical wavelet basis decompositions over which they exhibit visual improvement in terms of compression. This approach can be viewed as a bridge between edge detection and the nonlinear multiresolution representations of Ami Harten (see Journal of Applied Numerical Mathematics, vol.12, p.153-93, 1993).","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127733925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Segmentation of vector images by N-level-set-fitting","authors":"T. Hanning, H. Farr, M. Kellner, Verena Lauren","doi":"10.1109/ICIP.2001.958613","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958613","url":null,"abstract":"In many applications of segmentation algorithms the number of desired segments is known previously. We present a technique to segment a given vector image (in most cases consisting of three color channels) in a prior known number of segments consisting of connected pixel sets. The main idea is to minimize the Euclidean distance of a vector valued step function to the image, with the step function being constant on a segment. A local minimum of this optimization problem can be obtained by a simple merging algorithm, which starts with a segmentation of the image into a much greater number of segments. The starting segmentation can be computed by using well known histogram based thresholding algorithms.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127750035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Look-up-table based DCT domain inverse motion compensation","authors":"Shizhong Liu, A. Bovik","doi":"10.1109/ICIP.2001.958656","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958656","url":null,"abstract":"DCT-based digital video coding standards such as MPEG and H.26x have been widely adopted for multimedia applications. Thus video processing in the DCT domain usually proves to be more efficient than in the spatial domain. To directly convert an inter-coded frame into an intra-coded frame in the DCT domain, the problem of DCT domain inverse motion compensation was studied by Chang and Messerschmitt(1995). Since the data is organized block by block in the DCT domain, the DCT domain inverse motion compensation is computationally intensive. In this paper, a look-up-table (LUT) based method for DCT domain inverse motion compensation is proposed by modeling the statistical distribution of the DCT coefficients in typical images and video sequences. Compared to the method of Chang et al., the LUT based method can save more than 50% of the computing time based on experimental results. The memory requirement of the LUT is about 800 KB which is reasonable. Moreover, the LUT can be shared by multiple DCT domain video processing applications running on the same computer.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127889109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Analysis of LSB based image steganography techniques","authors":"R. Chandramouli, N. Memon","doi":"10.1109/ICIP.2001.958299","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958299","url":null,"abstract":"There have been many techniques for hiding messages in images in such a manner that the alterations made to the image are perceptually indiscernible. However, the question whether they result in images that are statistically indistinguishable from untampered images has not been adequately explored. We look at some specific image based steganography techniques and show that an observer can indeed distinguish between images carrying a hidden message and images which do not carry a message. We derive a closed form expression of the probability of detection and false alarm in terms of the number of bits that are hidden. This leads us to the notion of steganographic capacity, that is, how many bits can we hide in a message without causing statistically significant modifications? Our results are able to provide an upper bound on the this capacity. Our ongoing work relates to adaptive steganographic techniques that take explicit steps to foil the detection mechanisms. In this case we hope to show that the number of bits that can be embedded increases significantly.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126447325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
R. Molina, A. Katsaggelos, J. Mateos, C. A. Segall
{"title":"Bayesian high-resolution reconstruction of low-resolution compressed video","authors":"R. Molina, A. Katsaggelos, J. Mateos, C. A. Segall","doi":"10.1109/ICIP.2001.958415","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958415","url":null,"abstract":"A method for simultaneously estimating the high-resolution frames and the corresponding motion field from a compressed low-resolution video sequence is presented. The algorithm incorporates knowledge of the spatio-temporal correlation between low and high-resolution images to estimate the original high-resolution sequence from the degraded low-resolution observation. Information from the encoder is also exploited, including the transmitted motion vectors, quantization tables, coding modes and quantizer scale factors. Simulations illustrate an improvement in the peak signal-to-noise ratio when compared with traditional interpolation techniques and are corroborated with visual results.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126509466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Extended fuzzy rules for image segmentation","authors":"L. Dooley, G. Karmakar","doi":"10.1109/ICIP.2001.958319","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958319","url":null,"abstract":"The generic fuzzy rule-based image segmentation (GFRIS) technique does not produce good results for non-homogeneous regions that possess abrupt changes in pixel intensity, because it fails to consider two important properties of perceptual grouping, namely surroundedness and connectedness. A new technique called extended fuzzy rules for image segmentation (EFRIS) is proposed, which includes a second rule to that defined already in GFRIS, that incorporates both the surroundedness and connectedness properties of a region's pixels. This additional rule is based on a split-and-merge algorithm and refines the output from the GFRIS technique. Two different classes of image, namely light intensity and medical X-rays are empirically used to assess the performance of the new technique. Quantitative evaluation of the performance of EFRIS is discussed and contrasted with GFRIS using one of the standard segmentation evaluation methods. Overall, EFRIS exhibits significantly improved results compared with the GFRIS approach.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126509898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Separating geometry from texture to improve face analysis","authors":"Albert Pujol, J. L. Alba, J. Villanueva","doi":"10.1109/ICIP.2001.958583","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958583","url":null,"abstract":"This article studies the effect of preprocessing a classical PCA decomposition using a modified self organizing map (SOM) in order to find shape clusters to improve the texture analysis by means of a pool of PCAs. In most successful view-based recognition systems, shape and texture are jointly used to model statistically a linear or piece-wise linear subspace that optimally explains the face space for a specific database. Our work is aimed at separating the influence that variance in face shape stamps on the set of eigenfaces in the classical PCA decomposition. A set of experiments show the reliability of this new system.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"2013 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128089943","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Bernhard Fröba, T. Kastner, W. Zink, Christian Kiiblbeck
{"title":"Real-time active shape models for face segmentation","authors":"Bernhard Fröba, T. Kastner, W. Zink, Christian Kiiblbeck","doi":"10.1109/ICIP.2001.958989","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958989","url":null,"abstract":"In this work we tackle the problem of real-time alignment of active shape models to new object instances at video frame rate. To achieve this we use edge orientation information as the basic image feature. Unlike in the original active shape framework we incorporate the image features directly into the model vector. We also introduce a new update rule for a model point in a local surrounding. We demonstrate the effectiveness of the proposed approach with a face segmentation task. There we are able to fit a new model face on average within 20 msec on 500 MHz Pentium II PC if the initial model position and size does not deviate too much from the true position. The latter is assured by a face detection step which is carried out before the active shape alignment.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128142737","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A min-max approach to the multidimensional nonuniform FFT: application to tomographic image reconstruction","authors":"B. Sutton, J. Fessler","doi":"10.1109/ICIP.2001.959143","DOIUrl":"https://doi.org/10.1109/ICIP.2001.959143","url":null,"abstract":"The FFT is used widely in signal processing for efficient computation of the Fourier transform (FT) over a set of uniformly spaced frequency locations. However, in many applications, one requires nonuniform sampling in the frequency domain, i.e., a nonuniform FT. Several papers have described fast approximations for the nonuniform FT based on interpolating an oversampled FFT. This paper presents a method for the nonuniform FT that is optimal in a min-max sense. The proposed method minimizes the worst-case approximation error over all signals of unit norm. Unlike many previous methods for the nonuniform FT, the proposed method easily generalizes to multidimensional signals. We are investigating this method as a fast algorithm for computing the Radon transform in 2D iterative tomographic image reconstruction.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"22 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125685535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Face modeling for recognition","authors":"Rein-Lien Hsu, Anil K. Jain","doi":"10.1109/ICIP.2001.958588","DOIUrl":"https://doi.org/10.1109/ICIP.2001.958588","url":null,"abstract":"3D human face models have been widely used in applications such as facial animation, video compression/coding, augmented reality, head tracking, facial expression recognition, human action recognition, and face recognition. Modeling human faces provides a potential solution to identifying faces with variations in illumination, pose, and facial expression. We propose a method of modeling human faces based on a generic face model (a triangular mesh model) and individual facial measurements containing both shape and texture information. The modeling method adapts a generic face model to the given facial features, extracted from registered range and color images, in a global-to-local fashion. It iteratively moves the vertices of the mesh model to smoothen the non-feature areas, and uses the 2.5D active contours to refine feature boundaries. The resultant face model has been shown to be visually similar to the true face. Initial results show that the constructed model is quite useful for recognizing nonfrontal views.","PeriodicalId":291827,"journal":{"name":"Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115903735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}