{"title":"Using histograms to detect and track objects in color video","authors":"Michael Mason, Zoran Duric","doi":"10.1109/AIPR.2001.991219","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991219","url":null,"abstract":"Two methods of detecting and tracking objects in color video are presented. Color and edge histograms are explored as ways to model the background and foreground of a scene. The two types of methods are evaluated to determine their speed, accuracy and robustness. Histogram comparison techniques are used to compute similarity values that aid in identifying regions of interest. Foreground objects are detected and tracked by dividing each video frame into smaller regions (cells) and comparing the histogram of each cell to the background model. Results are presented for video sequences of human activity.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128671796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"High storage capacity architecture for pattern recognition using an array of Hopfield neural networks","authors":"Ming-Jung Seow, V. Asari","doi":"10.1109/AIPR.2001.991221","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991221","url":null,"abstract":"A new approach for the recognition of images using a two dimensional array of Hopfield neural networks is presented in this paper. In the proposed method, the N/spl times/N image is divided into sub-blocks of size M/spl times/M. Two-dimensional Hopfield neural networks of size M/spl times/M are used to learn and recognize the sub-images. All the N/sup 2//M/sup 2/ Hopfield modules are functioning independently and are capable of recognizing the corrupted image successfully when they work together. It is shown mathematically that the network system converges in all circumstances. The performance of the proposed technique is evaluated by applying it into various binary and gray scale images. The gray scale images are treated in a three-dimensional perspective by considering an 8-bit gray scale image as 8 independent binary images. Eight layers of binary networks are used for the recognition purpose. A Fuzzy-ART based neural network is used for the classification and labeling of the outputs in the Hopfield network. By employing the new approach, it can be seen that the storage capacity of the entire pattern recognition system would be increased to 2/sup n/ where n=N/sup 2//M/sup 2/. Experiments conducted on different images of various sizes have shown that the proposed network structure can learn and recognize images even with 30% noise. In addition, the number of iterations required for the convergence of the network is significantly reduced and the number of synaptic weights required for the entire architecture is reduced from N/sup 4/ to N/sup 2/M/sup 2/. The proposed network structure is suitable for building dedicated hardware to enable the pattern recognition in real-time due to the requirement of less number of registers to store synaptic weights and reduced number of interconnections between neurons.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115391259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A conversational paradigm for multimodal human interaction","authors":"Francis K. H. Quek","doi":"10.1109/AIPR.2001.991207","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991207","url":null,"abstract":"We present an alternative to the manipulative and semaphoric gesture recognition paradigms. Human multimodal communicative behaviors form a tightly integrated whole. We present a paradigm multimodal analysis in natural discourse based on a feature decompositive psycholinguistically derived model that permits us to access the underlying structure and intent of multimodal communicative discourse. We outline the psycholinguistics that drive our paradigm, the Catchment concept that facilitates our getting a computational handle on discourse entities, and summarize some approaches and results that realize the vision. We show examples of such discourse-structuring features as handedness, types of symmetry, gaze-at-interlocutor and hand 'origos'. Such analysis is an alternative to the 'recognition of one discrete gesture out of k stylized whole gesture models' paradigm.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127584786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Suitability of synthetic imagery for ATR evaluation","authors":"D. Meredith, C. Walters, C. Hoover","doi":"10.1109/AIPR.2001.991203","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991203","url":null,"abstract":"There is a commonly held belief among many who participate in ATR development and evaluation that a lack of data (imagery from tactical sensors) is a significant obstacle in the production and thorough evaluation of high performance ATR systems. Recognizing that it may be too costly or even impossible to collect a sufficient quantity of real data to adequately represent all ATR mission scenarios of interest, synthetically generated imagery may be one way to overcome this obstacle. Given the impact that synthetic imagery could potentially have on the development of our future weapon systems and on the performance of these systems in future conflicts, it is critical that this technology be validated prior to its use in the evaluation of ATR systems or in the training of operational ATRs. In an Office of Secretary of Defense (OSD) funded effort, the Night Vision and Electronic Sensors Directorate (NVESD) of the Communication and Electronics Command (CECOM) developed a plan for validation that is being executed.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121533374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
R. Srikanchana, Kun Huang, J. Xuan, M. Freedman, Y. Wang
{"title":"Mixture of principal axes registration for change analysis in computer-aided diagnosis","authors":"R. Srikanchana, Kun Huang, J. Xuan, M. Freedman, Y. Wang","doi":"10.1109/AIPR.2001.991199","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991199","url":null,"abstract":"Non-rigid image registration is a prerequisite for many medical image analysis applications, such as image fusion of multi-modality images and quantitative change analysis of a temporal sequence in computer-aided diagnosis. By establishing the point correspondence of the extracted feature points, it is possible to recover the deformation using nonlinear interpolation methods such as the thin-plate-spline approach. However, it is a difficulty task to establish an exact point correspondence due to the high complexity of the nonlinear deformation existing in medical images. In this paper, a mixture of principal axes registration (mPAR) method is proposed to resolve the correspondence problem through a neural computational approach. The novel feature of mPAR is to align two point sets without needing to establish an explicit point correspondence. Instead, it aligns the two point sets by minimizing the relative entropy between their probability distributions, resulting in a maximum likelihood estimate of the transformation matrix. The registration process consists of: (1) a finite mixture scheme to establish an improved point correspondence and (2) a multilayer perceptron (MLP) neural network to recover the nonlinear deformation. The neural computation for registration used a committee machine to obtain a mixture of piecewise rigid registrations, which gives a reliable point correspondence using multiple extracted objects in a finite mixture scheme. Then the MLP was used to determine the coefficients of a polynomial transform using extracted cross-points of elongated structures as control points. We have applied our mPAR method to a temporal sequence of mammograms from a single patient. The experimental results show that mPAR not only improves the accuracy of the point correspondence but also results in a desirable error-resilience property for control point selection errors.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125055581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A qualitative image reconstruction from an axial image sequence","authors":"Philippe Guermeur, E. Pissaloux","doi":"10.1109/AIPR.2001.991222","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991222","url":null,"abstract":"This paper presents a method to process axial monocular image sequences for mobile robot obstacle detection. We do not aim to achieve a complete scene reconstruction, but only to evaluate the time to collision and surface orientation useful for robot obstacle avoidance. Using a planar facet representation we first calculate formally the velocity field generated by the camera motion. The apparent deformations, in conjunction with a projective model, are then used in order to evaluate the scene apparent movement with a wide angle camera. In practice, we process separately the tangential and radial components of the apparent velocity vectors, using the epipolar constraint. Noise resistance is improved by integration using the Green's and Stoke's theorems which provide a link with surface moments. Experimental results on synthesis and real images of indoor scenes are given, and their validity is discussed Potential applications include visual navigation, obstacle detection, visual servoing, and object recognition.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133283283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A basic hand gesture control system for PC applications","authors":"C. Cohen, G. Beach, G. Foulk","doi":"10.1109/AIPR.2001.991206","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991206","url":null,"abstract":"We discuss the issues involved in controlling computer applications via gestures composed of both static symbols and dynamic motions. Each gesture is modeled from either static model information or a linear-in-parameters dynamic system. Recognition occurs in a real-time environment using a small amount of processing time and memory. We examine which gestures are appropriate, how the gestures can be recognized, and which commands the gestures should control. The tracking method is detailed, along with its use in providing coordinates for the gesture control a PowerPoint presentation.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130136265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A multiple perspective spectral approach to object detection","authors":"R. Bonneau","doi":"10.1109/AIPR.2001.991212","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991212","url":null,"abstract":"Many applications for detection of objects such as video analysis require that candidate objects be observed over a range of perspectives in 3 dimensional space. As a result we must have a robust model and detection process for these objects in order to accurately detect them through a range of geometric transformations. In order to keep our detection process computationally efficient, we use a compact multiresolution model to represent the range of geometric transformations possible in the object to be detected. Additionally, we form an integrated likelihood ratio detection statistic to optimize the detection performance over the entire space of targets being examined. To demonstrate the performance of this algorithm we apply our results to a compressed video sequence and show the improvement of our integrated three dimensional model as a function of model order.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125090160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"PUPILS-enabling a dialogue between the machine and the brain","authors":"R. Heishman, Zoran Duric, H. Wechsler","doi":"10.1109/AIPR.2001.991208","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991208","url":null,"abstract":"The human eye has been called the window to the soul. One component of the eye, the pupil, is considered by some psychologists to be the single involuntary indicator of cognitive activity in the human brain. Much research is currently directed toward eye tracking for the purpose of determining the focus of a subject's attention. This information is then used in various HCI applications (e.g., to operate a system GUI visually). Our efforts in this area are quite unique in that the goal is to use the pupil response as a measure of attentiveness to a particular visual task. We first discuss the role of the pupil in this capacity and the feasibility of a system designed to monitor and interpret pupil response relative to specific visual stimuli. We then demonstrate our approach in implementing a machine vision system (PUPILS-PUPil InterLocution System) that endeavors to gauge a subject's degree of attentiveness relative to a specific visual task. Finally, we present the results of initial experiments using PUPILS, discuss the contributions and shortcomings of this initial prototype and expound our plan for continued research in this area.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127177901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Directional edge registration for temporal chest image subtraction","authors":"Hui Zhao, S. Lo, M. Freedman","doi":"10.1109/AIPR.2001.991200","DOIUrl":"https://doi.org/10.1109/AIPR.2001.991200","url":null,"abstract":"We used a directional filtering technique to accurately register the ribs on temporal chest radiographs. Rib registration was the primary technical objective. In order to accurately extract the ribs, we developed a directional edge function that used rotating kernels through a dynamic search technique to extract rib edges of varying contrast. A directionally-weighted operator with a phase-contrast feature is embedded in the kernel design, which allows us to separate the upper and lower bounds of structures in the vertical direction. Broken rib edges were reconnected using a reasoning algorithm by analyzing the rib position and curve. The lung and heart boundaries were extracted using an inverted umbrella filter. Control points were designated at turning points and distributed on smooth edge segments on the temporal images. A thin-plate spline technique took the control points and performed the final registration prior to the temporal image subtraction. We found that normal chest structures (i.e. ribs, heart and normal lung structures) were greatly reduced in the subtraction images. In addition, the contrast of cancer spots was significantly increased.","PeriodicalId":277181,"journal":{"name":"Proceedings 30th Applied Imagery Pattern Recognition Workshop (AIPR 2001). Analysis and Understanding of Time Varying Imagery","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115216594","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}