{"title":"Adaptive photo collection page layout","authors":"C. B. Atkins","doi":"10.1109/ICIP.2004.1421718","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1421718","url":null,"abstract":"This paper presents a new photo collection page layout that attempts to maximize page coverage without having photos overlap. Layout is based on a hierarchical page partition, which provides explicit control over the aspect ratios and relative areas of the photos. We present an efficient method for finding a partition that produces a photo arrangement suitable for the shape of the page. Rather than relying on a stochastic search we employ a deterministic procedure that mimics the natural process of adding photos to the layout one by one.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"346 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115893385","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Omni-directional face detection based on real AdaBoost","authors":"Chang Huang, Bo Wu, H. Ai, S. Lao","doi":"10.1109/ICIP.2004.1418824","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1418824","url":null,"abstract":"We propose an omni-directional face detection method based on the confidence-rated AdaBoost algorithm, called real AdaBoost, proposed by R.E. Schapire and Y. Singer (see Machine Learning, vol.37, p.297-336, 1999). To use real AdaBoost, we configure the confidence-rated look-up-table (LUT) weak classifiers based on Haar-type features. A nesting-structured framework is developed to combine a series of boosted classifiers into an efficient object detector. For omni-directional face detection, our method has achieved a rather high performance and the processing speed can reach 217 ms per 320/spl times/240 image. Experiment results on the CMU+MIT frontal and the CMU profile face test sets are reported to show its effectiveness.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115958425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Medical image segmentation with minimal path deformable models","authors":"Pingkun Yan, A. Kassim","doi":"10.1109/ICIP.2004.1421669","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1421669","url":null,"abstract":"This paper presents an algorithm that segments medical images by extracting object contours. It delineates object boundaries by detecting a path with the minimum energy on the image. A worm algorithm based on deformable models is proposed to find the minimal path by using the dynamic programming technique. The proposed algorithm overcomes the shortcomings of traditional deformable models such as fastidious initialization and inefficiency on segmenting objects with complex shapes or topologies. After presenting the algorithm, its performance on various synthetic and medical images is shown. Experimental results indicate that the proposed algorithm is robust to noise and edge discontinuities.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132349097","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Rotation invariant texture features using rotated complex wavelet for content based image retrieval","authors":"M. Kokare, P. Biswas, B. N. Chatterji","doi":"10.1109/ICIP.2004.1418773","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1418773","url":null,"abstract":"A new rotationally invariant texture feature extraction method is introduced that utilizes the dual tree rotated complex wavelet filters (DT-RCWF) and dual tree complex wavelet transform (DT-CWT) jointly. A new two-dimensional rotated complex wavelet filter is designed with a complex wavelet filter coefficient. Decomposing the image with DT-RCWF and DT-CWT jointly gives shift invariant subbands oriented in twelve different directions. Isotropic rotationally invariant features are extracted from these subbands. The performance of image retrieval with the proposed features on rotated and nonrotated image databases is compared with the existing method. Experimental results show that the proposed rotation-invariant texture features are more robust and outperform the other existing methods.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132350057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
K. Sugimoto, Mitsuru Kobayashi, Yoshinori Suzuki, S. Kato, C. Boon
{"title":"Inter frame coding with template matching spatio-temporal prediction","authors":"K. Sugimoto, Mitsuru Kobayashi, Yoshinori Suzuki, S. Kato, C. Boon","doi":"10.1109/ICIP.2004.1418791","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1418791","url":null,"abstract":"A new algorithm is proposed for predicting pixels for inter frame coding without side information. There are many approaches in the past that exploited either spatial or temporal correlations for generating prediction signals of an image in a block-by-block basis. Our method proposed in this paper exploits both spatial and temporal correlations at once to predict the pixels to be encoded. The prediction is achieved by using a template matching mechanism, with reference to previously reconstructed groups of pixels in the same frame or adjacent frames, to fill in the pixels of target regions of a frame. This process is conducted at both the encoder and decoder, and hence allows the decoder to produce the same predictor as the encoder does without any side information. Our coder uses the proposed prediction in addition to conventional motion compensation means. Simulation results show that our approach achieves up to 11.14% of improvements at the same PSNR over a codec which uses conventional block-based motion compensation.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132565294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Optimal hierarchical representation and simulation of cloth and deformable objects","authors":"K. Moustakas, D. Tzovaras, M. Strintzis","doi":"10.1109/ICIP.2004.1421747","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1421747","url":null,"abstract":"This paper presents a novel pyramidal representation scheme for deformable object modelling, which uses a hierarchical approach to optimize the system's performance by executing the simulation in every level of a pyramid. The simulation results of each level are used to predict the lower level's state. The prediction is used as initial guess for the simulation of the lower level. The above procedure is repeated until the final level is reached. Experimental evaluation demonstrates that the proposed scheme is able to reduce significantly the computational cost, especially when the simulation involves procedures, which need large numerical computation like the conjugate gradient in implicit integration schemes.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130429446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
R. Bernardini, M. Durigon, R. Rinaldo, Luca Celetto, A. Vitali
{"title":"Polyphase spatial subsampling multiple description coding of video streams with H264","authors":"R. Bernardini, M. Durigon, R. Rinaldo, Luca Celetto, A. Vitali","doi":"10.1109/ICIP.2004.1421797","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1421797","url":null,"abstract":"In this work, we propose a multiple description (MD) coding system for video streams. In particular, our scheme originates four descriptions from the spatially downsampled polyphase components of the original frames. Each description is compressed independently with the recent H264/AVC video coding standard, it is packetized and sent over an error prone network. In case of errors in one or more descriptions, appropriate concealing is applied at the receiver, before insertion of the corrected frames into the corresponding receiver frame buffers. We propose and compare different concealment solutions and a post processing stage to attenuate visual effects related to MD coding. We analyze the trade off between robustness to channel errors and coding efficiency, comparing the proposed technique with single description (SD) video coding with H264/AVC. Experimental results validate the effectiveness of the proposed scheme.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"265 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133905204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Facial event mining using coupled hidden Markov models","authors":"Limin Ma, Qiang-feng Zhou, M. Celenk, D. Chelberg","doi":"10.1109/ICIP.2004.1419765","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1419765","url":null,"abstract":"Facial event mining is one of the key techniques for automatic human face analysis. It plays an important role in human computer interaction. This paper proposes a new approach to facial event recognition by combining active shape models (ASMs) and coupled hidden Markov models (CHMMs). Based on the assumption that a complex facial event can be decomposed into multiple coupled processes, ASMs are used to track global facial features and to decouple pattern attributes for upper and lower faces separately. These two interacting processes are modeled as a CHMM for training and recognition. Four basic facial events are investigated. Preliminary experiments yield consistent results that show the significant advantage of CHMMs over conventional HMMs for facial event mining in video.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133946222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust reconstruction of 3D points from images","authors":"R. Rodrigues, A. Fernandes","doi":"10.1109/ICIP.2004.1421748","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1421748","url":null,"abstract":"This paper presents a robust approach for 3D point reconstruction based on a set of images taken from a static scene with known, but not necessarily exact or regular, camera parameters. The points to be reconstructed are chosen from the contours of images, and a world-based formulation of the reconstruction problem and associated epipolar geometry is used. The result is a powerful mean of transparently integrating contributions from multiple images, and increased robustness to situations such as occlusions or apparent contours. Two steps for adding robustness are proposed: cross-checking, which validates a reconstructed point taken from an image by projecting it on a special subset of the remaining images; and merging, which fuses pairs of reconstructed points that are close in 3D space and that were initially chosen from different images. Results obtained with a synthetic scene (for ground truth comparison and error assessment), and two real scenes show the improved robustness achieved with the steps proposed.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131622268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Image fusion based on nonnegative matrix factorization","authors":"Junying Zhang, Le Wei, Q. Miao, Y. Wang","doi":"10.1109/ICIP.2004.1419463","DOIUrl":"https://doi.org/10.1109/ICIP.2004.1419463","url":null,"abstract":"Nonnegative Matrix Factorization technique (NMF) has been shown to have various applications to image processing, because of its power of local or part-based representation of objects and/or images. In this paper, we present an image fusion method based on NMF, not by the part-based representation feature of NMF, but by its wholly representation of the images needed to be fused: the images are fused by NMF with the parameter r of the NMF to be set to 1. Our experimental results show that the proposed method is efficient and effective for image fusion compared with many other image fusion methods.","PeriodicalId":184798,"journal":{"name":"2004 International Conference on Image Processing, 2004. ICIP '04.","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131655894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}