Jiali Cui, Yunhong Wang, Junzhou Huang, T. Tan, Zhenan Sun
{"title":"An iris image synthesis method based on PCA and super-resolution","authors":"Jiali Cui, Yunhong Wang, Junzhou Huang, T. Tan, Zhenan Sun","doi":"10.1109/ICPR.2004.1333804","DOIUrl":"https://doi.org/10.1109/ICPR.2004.1333804","url":null,"abstract":"It is very important for the performance evaluation of iris recognition algorithms to construct very large iris databases. However, limited by the real conditions, there are no very large common iris databases now. In this paper, an iris image synthesis method based on principal component analysis (PCA) and super-resolution is proposed. The iris recognition algorithm based on PCA is first introduced and then, iris image synthesis method is presented. The synthesis method first constructs coarse iris images with the given coefficients. Then, synthesized iris images are enhanced using super-resolution. Through controlling the coefficients, we can create many iris images with specified classes. Extensive experiments show that the synthesized iris images have satisfactory cluster and the synthesized iris databases can be very large.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114588865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Reducing artifacts in BDCT-coded images by adaptive pixel-adjustment","authors":"J. Zou","doi":"10.1109/ICPR.2004.1334179","DOIUrl":"https://doi.org/10.1109/ICPR.2004.1334179","url":null,"abstract":"Generating artifacts is a major problem with compressed images coded by the block discrete cosine transform (BDCT). An iterative deblocking method based on projection onto convex sets (POCS) is presented in this paper. The method considers pixel-wise local adaptability. The adjustment of a pixel is determined by local properties of the pixel. A coded image is iteratively projected onto two locally adaptive smoothness constraint sets, a locally adaptive quantization constraint set and an intensity constraint set to obtain an improved image where artifacts are reduced. The proposed method is tested on JPEG encoded images with excellent results.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114609782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sanae Shimizu, Kazuhiko Yamamoto, Caihua Wang, Yutaka Sato, H. Tanahashi, Y. Niwa
{"title":"Moving object detection with mobile stereo omni-directional system (SOS) based on motion compensatory inter-frame depth subtraction","authors":"Sanae Shimizu, Kazuhiko Yamamoto, Caihua Wang, Yutaka Sato, H. Tanahashi, Y. Niwa","doi":"10.1109/ICPR.2004.1334514","DOIUrl":"https://doi.org/10.1109/ICPR.2004.1334514","url":null,"abstract":"Moving object detection with a mobile image sensor is an important task when considering mobile robots for use in human environments. In this paper, we propose a novel method/or effectively solving the problem of detecting moving objects for mobile robots by using the stereo omni-directional system (SOS) which has a complete spherical FOV. We first predict the depth image for the present time from the self-motion of the SOS and the depth image obtained at the previous time, and then detect the moving objects by comparing the predicted depth image with the actual one obtained at the present time. Experiments in the real world show the effectiveness of the proposed method.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116618000","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Discovery of the tri-edge inequality with binary vector dissimilarity measures","authors":"Bin Zhang, S. Srihari","doi":"10.1109/ICPR.2004.1333861","DOIUrl":"https://doi.org/10.1109/ICPR.2004.1333861","url":null,"abstract":"In certain spaces using some distance measures, the sum of any two distances is always bigger than the third one. Such a special property is called the tri-edge inequality (TEI). In this paper, the tri-edge inequality characterizing several binary distance measures is mathematically proven and experimentally verified, and the implications of TEI are discussed as well.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117261073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Face detection using discriminating feature analysis and support vector machine in video","authors":"P. Shih, Chengjun Liu","doi":"10.1109/ICPR.2004.1334236","DOIUrl":"https://doi.org/10.1109/ICPR.2004.1334236","url":null,"abstract":"This work presents a novel face detection method in video by using discriminating feature analysis (DFA) and support vector machine (SVM). Our method first incorporates temporal and skin color information to locate the field of interests. Then the face class is modelled using a small training set and the nonface class is defined by choosing nonface images that lie close to the face class. Finally, the SVM classifier together with Bayesian statistical analysis procedure applies the efficient features defined by DFA for face and nonface classification. Experiments using both still images and video streams show the feasibility of our new face detection method. In particular, when using 92 images (containing 282 faces) from the MIT-CMU test sets, our method achieves 98.2% correct face detection accuracy with 2 false detections. When using video streams, our method detects faces reliably with computational efficiency of more than 20 frames per second.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"80 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128411331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust direction estimation of gradient vector field for iris recognition","authors":"Zhenan Sun, Yunhong Wang, T. Tan, Jiali Cui","doi":"10.1109/ICPR.2004.1334375","DOIUrl":"https://doi.org/10.1109/ICPR.2004.1334375","url":null,"abstract":"As a reliable personal identification method, iris recognition has been receiving increasing attention. Based on the theory of robust statistics, a novel geometry-driven method for iris recognition is presented in this paper. An iris image is considered as a 3D surface of piecewise smooth patches. The direction of the 2D vector, which is the planar projection of the normal vector of image surface, is illumination insensitive and opposite to the direction of gradient vector. So the directional information of iris image's gradient vector field (GVF) is used to represent iris pattern. Robust direction estimation, direction diffusion followed by vector directional filtering, is performed on the GVF to extract stable iris feature. Extensive experimental results demonstrate that the recognition performance of the proposed algorithm is comparable with the best method in the open literature.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"126 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128454877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Object-based and event-based semantic video adaptation","authors":"M. Bertini, A. Bimbo, R. Cucchiara, A. Prati","doi":"10.1109/ICPR.2004.640","DOIUrl":"https://doi.org/10.1109/ICPR.2004.640","url":null,"abstract":"Semantic video adaptation allows transmitting video content with different viewing quality, depending on the relevance of the content from the user's viewpoint. To this end, an automatic annotation subsystem must be employed that automatically detect relevant objects and events in the video stream. We present a composite framework that is made of an automatic annotation engine and a semantics-based adaptation module. Three new different compression solutions are proposed that work at the object or event level. Their performance is compared according to a new measure that takes into account the user's satisfaction and the effects on it of the errors in the annotation module.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128701830","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Address-block extraction by Bayesian rule","authors":"T. Kagehiro, Masashi Koga, H. Sako, H. Fujisawa","doi":"10.1109/ICPR.2004.1334315","DOIUrl":"https://doi.org/10.1109/ICPR.2004.1334315","url":null,"abstract":"A method for extracting a recipient address-block from a mail image has been developed. The method is composed of two steps: nomination of address-block candidates and evaluation of these candidates by using the Bayesian rule according to each of address-block type. Accordingly, the proposed method can cope with various types of address-blocks. The effectiveness of the method was confirmed in several address extraction experiments. These experiments show that the top-five extraction results include one correct address-block in 94% of total number of printed-mail cases and 89% in of handwritten-mail cases.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129070158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A miniature stereo vision machine (MSVM-III) for dense disparity mapping","authors":"Yunde Jia, Xiaoxun Zhang, Mingxiang Li, L. An","doi":"10.1109/ICPR.2004.1334290","DOIUrl":"https://doi.org/10.1109/ICPR.2004.1334290","url":null,"abstract":"We have developed a miniature stereo vision machine (MSVM-III) with three cameras for generating high-resolution dense disparity maps at the video rate. The MSVM-III uses only one FPGA chip to compactly compute trinocular rectification, LoG filtering, and area-based matching. The machine, running at 60 MHz, could process more than 30 fps dense disparity maps with 640/spl times/480 pixels in 64-pixel disparity search range, and 120 fps with 320/spl times/240 pixels. Moreover, the MSVM-III has an IEEE 1394 port to a host at the video rate, an interface port to LCD as a miniature 3D imager, and a user board for controlling small mobile robot or other autonomous systems.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124627383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Sketched symbol recognition using Zernike moments","authors":"H. Hse, A. Newton","doi":"10.1109/ICPR.2004.1334128","DOIUrl":"https://doi.org/10.1109/ICPR.2004.1334128","url":null,"abstract":"We present an on-line recognition method for hand-sketched symbols. The method is independent of stroke-order, -number, and -direction, as well as invariant to scaling, translation, rotation and reflection of symbols. Zernike moment descriptors are used to represent symbols and three different classification techniques are compared: support vector machines (SVM), minimum mean distance (MMD), and nearest neighbor (NN). We have obtained a 97% recognition accuracy rate on a dataset consisting of 7,410 sketched symbols using Zernike moment features and a SVM classifier.","PeriodicalId":335842,"journal":{"name":"Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004.","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129433047","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}