{"title":"Elliptic Metric K-NN Method with Asymptotic MDL Measure","authors":"T. Satonaka, K. Uchimura","doi":"10.1109/ICIP.2006.312864","DOIUrl":"https://doi.org/10.1109/ICIP.2006.312864","url":null,"abstract":"We describe an adaptive metric learning model combining the generative and the discriminative models for the face recognition. The asymptotic model based on the MDL measure is formulated for each class to estimate the variance by using small training examples. The feature fusion method is introduced to assume the missing patterns between the classes and to deal with the k-th nearest neighbor classification. The metric parameters obtained from the asymptotic MDL estimation are refined by using the synthesized feature patterns. We demonstrate an improved recognition performance on the ORL and UMIST face databases.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115716591","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Disparity-Based 3D Face Modeling for 3D Face Recognition","authors":"A. Ansari, M. Abdel-Mottaleb, M. Mahoor","doi":"10.1109/ICIP.2006.312416","DOIUrl":"https://doi.org/10.1109/ICIP.2006.312416","url":null,"abstract":"We present an automatic disparity-based approach for 3D face modeling, from two frontal and one profile view stereo images, for 3D face recognition applications. Once the images are captured, the algorithm starts by extracting selected 2D facial features from one of the frontal views and computes a dense disparity map from the two frontal images. We then align a low resolution 2D mesh model to the selected features, adjust some of its vertices along the profile line using the profile view, increase its triangular vertices to a higher resolution, and re-project them back on the frontal image. Using the coordinates of the re-projected vertices and their corresponding disparities, we capture and compute the 3D facial shape variations using stereo vision. The final result is a deformed 3D model specific to a given subject's face. Application of the model in 3D face recognition validates the algorithm and shows a promising 98 % recognition rate.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125567293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Martin Luessi, Marco Eichmann, G. Schuster, A. Katsaggelos
{"title":"New Results on Efficient Optimal Multilevel Image Thresholding","authors":"Martin Luessi, Marco Eichmann, G. Schuster, A. Katsaggelos","doi":"10.1109/ICIP.2006.312426","DOIUrl":"https://doi.org/10.1109/ICIP.2006.312426","url":null,"abstract":"Image thresholding is one of the most common image processing operations, since almost all image processing schemes need some sort of separation of the pixels into different classes. In order to find the thresholds, almost all methods analyze the histogram of the image. In most cases, the optimal thresholds are found by either minimizing or maximizing an objective function, which depends on the positions of the thresholds. We identify two classes of objective functions for which the optimal thresholds can be found by algorithms with low time complexity. We show, that for example the method proposed by Otsu (1979) and other well known methods have objective functions belonging to these classes. By implementing the algorithms in ANSI C and comparing their execution times, we can make a quantitative statement about their performance.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114275994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Guanglei Xiong, Xiaobo Zhou, L. Ji, P. Bradley, N. Perrimon, Stephen T. C. Wong
{"title":"Segmentation of Drosophila RNAI Fluorescence Images Using Level Sets","authors":"Guanglei Xiong, Xiaobo Zhou, L. Ji, P. Bradley, N. Perrimon, Stephen T. C. Wong","doi":"10.1109/ICIP.2006.312365","DOIUrl":"https://doi.org/10.1109/ICIP.2006.312365","url":null,"abstract":"Image-based, high throughput genome-wide RNA interference (RNAi) experiments are increasingly carried out to facilitate the understanding of gene functions in intricate biological processes. Robust automated segmentation of the large volumes of output images generated from image-based screening is much needed for data analyses. In this paper, we propose a new automated segmentation technique to fill the void. The technique consists of two steps: nuclei and cytoplasm segmentation. In the former step, nuclei are extracted, labeled and used as starting points for the latter. A new force obtained from rough segmentation is introduced into the classical level set curve evolution to improve the performance for odd shapes, such as spiky or ruffly cells. A scheme of preventing curves from crossing is proposed to treat the difficulty of segmenting touching cells. We apply it to three types of drosophila cells in RNAi fluorescence images. In all cases, greater than 92% accuracy is obtained.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"2014 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129906803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. Tu, Thomas S. Huang, Yingen Xiong, R. Rose, Francis K. H. Quek
{"title":"Calibrating Head Pose Estimation in Videos for Meeting Room Event Analysis","authors":"J. Tu, Thomas S. Huang, Yingen Xiong, R. Rose, Francis K. H. Quek","doi":"10.1109/ICIP.2006.313066","DOIUrl":"https://doi.org/10.1109/ICIP.2006.313066","url":null,"abstract":"In this paper, we study the calibration of head pose estimation in stereo camera setting for meeting room video event analysis. Head pose information infers the direction of attention of the subjects in video, therefore is valuable for video event analysis/indexing, especially in meeting room scenario. We are developing a multi-modal meeting room data analyzing system for studying meeting room interaction dynamics, in which head pose estimation is one of the key components. As each subject in the meeting room can be observed by a pair of stereo cameras, we do 2D head tracking for the subject in each camera, and the 3D coordinate of the head can be obtained by triangulation. The 3D head pose is estimated in one of the camera coordinate system, we develop a procedure to accurately convert the estimated 3D pose in the camera coordinate system to that in the world coordinate system. In the experiment, visualization of the estimated head pose and location in world coordinate system verifies the soundness of our design. The estimated head pose and 3D location of the subjects in the meeting room allows further analysis of meeting room interaction dynamics, such as F-formation, floor-control, etc.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129102104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Model-Based Evaluation Methodology for Assessing the Efficacy of Packet-Level FEC for Delay-Constrained Video Network Transport","authors":"Xunqi Yu, J. Modestino, Y. Chan","doi":"10.1109/ICIP.2006.312554","DOIUrl":"https://doi.org/10.1109/ICIP.2006.312554","url":null,"abstract":"We propose an analytical model to study the overall effectiveness of packet-level FEC in improving delay-constrained video transport over wired networks subject to random loss and delay. In particular, we provide an exact analysis for the effect of the additional delay resulting from FEC coding/ decoding on end-to-end video transmission quality. The effects of several key system parameters on the effectiveness of FEC are investigated and the tradeoffs associated with choice of these parameters are demonstrated by some specific numerical examples.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123530233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Perceptual Feature Selection for Semantic Image Classification","authors":"Dejan Depalov, T. Pappas, Dongge Li, B. Gandhi","doi":"10.1109/ICIP.2006.313130","DOIUrl":"https://doi.org/10.1109/ICIP.2006.313130","url":null,"abstract":"Content-based image retrieval has become an indispensable tool for managing the rapidly growing collections of digital images. The goal is to organize the contents semantically, according to meaningful categories. In recent papers we introduced a new approach for semantic image classification that relies on the adaptive perceptual color-texture segmentation algorithm proposed by Chen et al. This algorithm combines knowledge of human perception and signal characteristics to segment natural scenes into perceptually uniform regions. The resulting segments can be classified into semantic categories using region-wide features as medium level descriptors. Such descriptors are the key to bridging the gap between low-level image primitives and high-level image semantics. The segment classification is based on linear discriminant analysis techniques. In this paper, we examine the classification performance (precision and recall rates) when different sets of region-wide features are used. These include different color composition features, spatial texture, and segment location. We demonstrate the effectiveness of the proposed techniques on a database that includes 9000 segments from approximately 2500 photographs of natural scenes.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121287700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Y. Fougerolle, A. Gribok, S. Foufou, F. Truchetet, M. Abidi
{"title":"Supershape Recovery from 3D Data Sets","authors":"Y. Fougerolle, A. Gribok, S. Foufou, F. Truchetet, M. Abidi","doi":"10.1109/ICIP.2006.312975","DOIUrl":"https://doi.org/10.1109/ICIP.2006.312975","url":null,"abstract":"In this paper, we apply supershapes and R-functions to surface recovery from 3D data sets. Individual supershapes are separately recovered from a segmented mesh. R-functions are used to perform Boolean operations between the reconstructed parts to obtain a single implicit equation of the reconstructed object that is used to define a global error reconstruction function. We present surface recovery results ranging from single synthetic data to real complex objects involving the composition of several supershapes and holes.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124337517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A New Contourlet Transform with Sharp Frequency Localization","authors":"Yue M. Lu, M. Do","doi":"10.1109/ICIP.2006.312657","DOIUrl":"https://doi.org/10.1109/ICIP.2006.312657","url":null,"abstract":"The contourlet transform was proposed as a directional multiresolution image representation that can efficiently capture and represent singularities along smooth object boundaries in natural images. Its efficient filter bank construction as well as low redundancy make it an attractive computational framework for various image processing applications. However, a major drawback of the original contourlet construction is that its basis images are not localized in the frequency domain. In this paper, we analyze the cause of this problem, and propose a new contourlet construction as a solution. Instead of using the Laplacian pyramid, we employ a new multiscale decomposition defined in the frequency domain. The resulting basis images are sharply localized in the frequency domain and exhibit smoothness along their main ridges in the spatial domain. Numerical experiments on image denoising show that the proposed new contourlet transform can significantly outperform the original transform both in terms of PSNR (by several dB 's) and in visual quality, while with similar computational complexity.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133645727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Parameter Estimation in Bayesian Reconstruction of Multispectral Images using Super Resolution Techniques","authors":"R. Molina, M. Vega, J. Mateos, A. Katsaggelos","doi":"10.1109/ICIP.2006.312720","DOIUrl":"https://doi.org/10.1109/ICIP.2006.312720","url":null,"abstract":"In this paper we present a new super resolution Bayesian method for pansharpening of multispectral images which: a) incorporates prior knowledge on the expected characteristics of the multispectral images, b) uses the sensor characteristics to model the observation process of both panchromatic and multispectral images, and c) performs the estimation of all the unknown parameters in the model. Using real data, the pansharpened multispectral images are compared with the images obtained by other pansharpening methods and their quality is assessed both qualitatively and quantitatively.","PeriodicalId":299355,"journal":{"name":"2006 International Conference on Image Processing","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132678080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}