{"title":"Optimizing image segmentation using color model mixtures","authors":"Aristide C. Chikando, J. Kinser","doi":"10.1109/AIPR.2005.38","DOIUrl":"https://doi.org/10.1109/AIPR.2005.38","url":null,"abstract":"Several mathematical color models have been proposed to segment images based on their color information content. The most frequently used color models of such sort include RGB, HSV, YCbCr, etc. These models were designed to represent color and in some cases emulate how the reflection of light on a given entity is perceived by the human eye. They were, however, not designed specifically for the purpose of image segmentation. In this study, the efficiency of several color models for the application of image segmentation is assessed and more efficient color models, consisting of color model mixtures, are explored. It was observed that two of the studied models, YCbCr and linear, were more efficient for the purpose of image segmentation. Additionally, by employing multivariate analysis, it was observed that the model mixtures were more efficient than the most commonly used models studied, and thus optimized the segmentation","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"189 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123514724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D scene modeling using sensor fusion with laser range finder and image sensor","authors":"Yunqian Ma, Z. Wang, Michael E. Bazakos, W. Au","doi":"10.1109/AIPR.2005.5","DOIUrl":"https://doi.org/10.1109/AIPR.2005.5","url":null,"abstract":"Activity detection (e.g. recognizing people's behavior and intent), when used over an extended range of applications, suffers from high false detection rates. Also, activity detection limited to 2D image domain (symbolic space) is confined to qualitative activities. Symbolic features, represented by apparent dimensions, i.e. pixels, can vary with distance or viewing angle. One way to enhance performance is to work within the physical space, where object features are represented by their physical dimensions (e.g. inches or centimeters) and are invariant to distance or viewing angle. In this paper, we propose an approach to construct a 3D site model and co-register the video with the site model to obtain real-time physical reference at every pixel in the video. We present a unique approach that creates a 3D site model via fusion of laser range sensor and a single camera. We present experimental results to demonstrate our approach.","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"105 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116117812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Medical image watermarking for multiple modalities","authors":"A. Maeder, B. Planitz","doi":"10.1109/AIPR.2005.33","DOIUrl":"https://doi.org/10.1109/AIPR.2005.33","url":null,"abstract":"Transfer of digital medical images between multiple parties requires the assurance of image identity and integrity, which can be achieved through image watermarking. This raises concerns for loss in viewer performance due to degradation of image quality. Here we describe an approach to ensure that impact on the image quality is well below the threshold of visual perceptibility. The principles on which this approach rests are the choice of a suitably light payload, and the use of different watermarking methods and parameters for different medical image types. We provide examples of this approach applied to MR, CT and CR images","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123709186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multimodal biometric identification for large user population using fingerprint, face and iris recognition","authors":"Teddy Ko","doi":"10.1109/AIPR.2005.35","DOIUrl":"https://doi.org/10.1109/AIPR.2005.35","url":null,"abstract":"Biometric systems based solely on one-modal biometrics are often not able to meet the desired performance requirements for large user population applications, due to problems such as noisy data, intra-class variations, restricted degrees of freedom, nonuniversity, spoof attacks, and unacceptable error rates. Multimodal biometrics refers to the use of a combination of two or more biometric modalities in a single identification system. The most compelling reason to combine different modalities is to improve the recognition accuracy. This can be done when features of different biometrics are statistically independent. This paper overviews and discusses the various scenarios that are possible in multimodal biometric systems using fingerprint, face and iris recognition, the levels of fusion that are possible and the integration strategies that can be adopted to fuse information and improve overall system accuracy. This paper also discusses how the image quality of fingerprint, face and iris used in the multimodal biometric systems affects the overall identification accuracy and the need of staffing for the secondary human validation. For a large user population identification system, which often has more than tens or hundreds of millions of subject images already enrolled in the matcher databases and has to process more than hundreds of thousands of identification requests, the system's identification accuracy and the need of staffing levels to properly operate the system are two of the most important factors in determining whether a system is properly designed and integrated","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121855627","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A control theoretic method for categorizing visual imagery as human motion behaviors","authors":"C. Cohen","doi":"10.1109/AIPR.2005.6","DOIUrl":"https://doi.org/10.1109/AIPR.2005.6","url":null,"abstract":"We propose a method that not only identifies humans in the environment and their location, but can also classify and identify their activity, providing a threat assessment. Such assessments would be useful for both human and vehicle activities in crowds to determine aberrant behavior from previously identified truth data sets. Such aberrant behavior would lead to IED detection, RPG detection, and the recognition of suicide bombers, before the explosives and planted and activated. The heuristics needed involve recognition of information bearing features in the environment, and the determination of how those features relate to each other over time (that is, gesture recognition). This paper addresses the mathematical development necessary to create a behavior and gait recognition sensor system that has its foundation on the recognition of combined individual gestures","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126583576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Girish Gopalakrishnan, S. Kumar, A. Narayanan, R. Mullick
{"title":"A fast piecewise deformable method for multi-modality image registration","authors":"Girish Gopalakrishnan, S. Kumar, A. Narayanan, R. Mullick","doi":"10.1109/AIPR.2005.7","DOIUrl":"https://doi.org/10.1109/AIPR.2005.7","url":null,"abstract":"Medical image fusion is becoming increasingly popular for enhancing diagnostic accuracy by intelligently 'fusing' information obtained from two different images. These images may be obtained from the same modality at different time instances or from multiple modalities recording complementary information. Due to the nature of the human body and also due to patient motion and breathing, there is a need for deformable registration algorithms in medical imaging. Typical nonparametric (deformable) registration algorithms such as the fluid-based, demons and curvature-based techniques are computationally intensive and have been demonstrated for mono-modality registrations only. We propose a fast and deformable algorithm using a 2-tiered strategy wherein a global MI-based affine registration is followed by a local piecewise refinement. We have tested this method on CT and PET images and validated the same using clinical experts","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132900749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Mahfouz, A. Fathy, Yunqiang Yang, Emam ElHak Ali, A. Badawi
{"title":"See-through-wall imaging using ultra wideband pulse systems","authors":"M. Mahfouz, A. Fathy, Yunqiang Yang, Emam ElHak Ali, A. Badawi","doi":"10.1109/AIPR.2005.40","DOIUrl":"https://doi.org/10.1109/AIPR.2005.40","url":null,"abstract":"Surveillance/navigation systems presently used make extensive use of television, infrared, and other line-of-sight-surveillance hardware. However, these systems cannot tell what is happening or locate persons/assets on the other side of a wall, behind bushes, in the dark, in a tunnel or a cave, or through a dense fog. It is our objective here to develop a new sensor, based on UWB technology. A small, lightweight, low power transceiver or multiples that are based upon the fact that microwave frequencies can be optimized to penetrate nonmetallic materials, and providing very precise ranging information. This new surveillance/navigation capability can help provide information about what is in a wall or on the other side of a door, and can be extended to provide precise global position in areas where these services are denied such as in tunnels or caves. This paper presents our efforts along these lines including image enhancements.","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133449341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An illuminance-reflectance nonlinear video enhancement model for homeland security applications","authors":"Li Tao, R. Tompkins, V. Asari","doi":"10.1109/AIPR.2005.14","DOIUrl":"https://doi.org/10.1109/AIPR.2005.14","url":null,"abstract":"A illuminance-reflectance model based video stream enhancement algorithm is proposed for improving the visual quality of digital video streams captured by surveillance camera under insufficient and/or nonuniform lighting conditions. The paper presents computational methods for estimation of scene illuminance and reflectance, adaptive dynamic range compression of illuminance, and adaptive enhancement for mid-tone frequency components. The images are processed in a similar way as human eyes sensing a scene. The algorithm demonstrates high quality of enhanced images, robust performance and fast processing speed. Compared with Retinex and multi-scale retinex with color restoration (MSRCR), the proposed method shows a better balance between luminance enhancement and contrast enhancement as well as a more consistent and reliable color rendition without introducing incorrect colors. This is an effective technique for image enhancement with simple computational procedures, which makes real-time enhancement for homeland security application successfully realized. The application of this image enhancement technique to the FRGC images yields improved face recognition results","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131199132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Discretization error based mesh generation for diffuse optical tomography","authors":"M. Guven, B. Yazıcı, Kiwoon Kwon, E. Giladi","doi":"10.1109/AIPR.2005.26","DOIUrl":"https://doi.org/10.1109/AIPR.2005.26","url":null,"abstract":"In this paper, we analyze the perturbation in the reconstructed optical absorption images, resulting from the discretization of the forward and inverse problems. We show that the perturbation due to each problem is a function of both the forward and inverse problem solutions and can be reduced by proper refinement of the discretization mesh. Based on the perturbation analysis, we devise an adaptive discretization scheme for forward and inverse problems, which reduces the perturbation on the reconstructed image. Such a discretization scheme leads to an adaptively refined composite mesh sufficient to approximate the forward and inverse problem solutions within a desired level of accuracy while keeping the computational complexity within the computational power limits","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131308441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Grouping sensory primitives for object recognition and tracking","authors":"R. Madhavan, Mike Foedisch, Tommy Chang, T. Hong","doi":"10.1109/AIPR.2005.29","DOIUrl":"https://doi.org/10.1109/AIPR.2005.29","url":null,"abstract":"In this paper, we describe our recent efforts in grouping sensory data into meaningful entities. Our grouping philosophy is based on perceptual organization principles using gestalt hypotheses where we impose structural regularity on sensory primitives stemming from a common underlying cause. We present results using field data from UGVs and outline the utility of our research in object recognition and tracking for autonomous vehicle navigation. In addition, we show how the grouping efforts can be useful for constructing symbolic topological maps when data from different sensing modalities are fused in a bottom-up and top-down fashion","PeriodicalId":130204,"journal":{"name":"34th Applied Imagery and Pattern Recognition Workshop (AIPR'05)","volume":"181 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132928580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}