{"title":"Hand gesture recognition following the dynamics of a topology-preserving network","authors":"F. Flórez, J. García, J. García, A. Hernandez","doi":"10.1109/AFGR.2002.1004173","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004173","url":null,"abstract":"We present a new structure capable of characterizing hand posture, as well as its movement. Topology of a self-organizing neural network determines posture, whereas its adaptation dynamics throughout time determines gesture. This adaptive character of the network allows us to avoid the correspondence problem of other methods, so that the gestures are modelled by the movement of the neurons. To validate this method, we have trained the system with 12 gestures, some of which are very similar, and have obtained high success rates (over 97%). This application of a self-organizing network opens up a new field of research because its topology is used to characterize the objects and not to classify them, as is usually the case.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"2016 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127367672","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Real-time multi-view face detection","authors":"ZhenQiu Zhang, Long Zhu, S. Li, HongJiang Zhang","doi":"10.1109/AFGR.2002.1004147","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004147","url":null,"abstract":"We present a detector-pyramid architecture for real-time multi-view face detection. Using a coarse to fine strategy, the full view is partitioned into finer and finer views. Each face detector in the pyramid detects faces of its respective view range. Its training is performed by using a new meta booting learning algorithm. This results in the first real-time multi-view face detection system which runs at 5 frames per second for 320/spl times/240 image sequence.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126740601","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A comparison of Gabor filter methods for automatic detection of facial landmarks","authors":"Ian R. Fasel, M. Bartlett, J. Movellan","doi":"10.1109/AFGR.2002.1004161","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004161","url":null,"abstract":"This paper presents a systematic analysis of Gabor filter banks for detection of facial landmarks (pupils and philtrum). Sensitivity is assessed using the A' statistic, a non-parametric estimate of sensitivity independent of bias commonly used in the psychophysical literature. We find that current Gabor filter bank systems are overly complex. Performance can be greatly improved by reducing the number of frequency and orientation components in these systems. With a single frequency band, we obtained performances significantly better than those achievable with current systems that use multiple frequency bands. Best performance for pupil detection was obtained with filter banks peaking at 4 iris widths per cycle and 8 orientations. Best performance for philtrum location was achieved with filter banks with 5.5 iris widths per circle and 8 orientations.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123200012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Real-time tracking of multiple fingertips and gesture recognition for augmented desk interface systems","authors":"Kenji Oka, Yoichi Sato, H. Koike","doi":"10.1109/AFGR.2002.1004191","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004191","url":null,"abstract":"We propose a fast and robust method for tracking a user's hand and multiple fingertips; we then demonstrate gesture recognition based on measured fingertip trajectories for augmented desk interface systems. Our tracking method is capable of tracking multiple fingertips in a reliable manner even in a complex background under a dynamically changing lighting condition without any markers. First, based on its geometrical features, the location of each fingertip is located in each input infrared image frame. Then, correspondences of detected fingertips between successive image frames are determined based on a prediction technique. Our gesture recognition system is particularly advantageous for human-computer interaction (HCI) in that users can achieve interactions based on symbolic gestures at the same time that they perform direct manipulation with their own hands and fingers. The effectiveness of our proposed method has been successfully demonstrated via a number of experiments.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121224239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Visual prosody: facial movements accompanying speech","authors":"H. Graf, E. Cosatto, V. Strom, Fu Jie Huang","doi":"10.1109/AFGR.2002.1004186","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004186","url":null,"abstract":"As we articulate speech, we usually move the head and exhibit various facial expressions. This visual aspect of speech aids understanding and helps communicating additional information, such as the speaker's mood. We analyze quantitatively head and facial movements that accompany speech and investigate how they relate to the text's prosodic structure. We recorded several hours of speech and measured the locations of the speakers' main facial features as well as their head poses. The text was evaluated with a prosody prediction tool, identifying phrase boundaries and pitch accents. Characteristic for most speakers are simple motion patterns that are repeatedly applied in synchrony with the main prosodic events. Direction and strength of head movements vary widely from one speaker to another, yet their timing is typically well synchronized with the spoken text. Understanding quantitatively the correlations between head movements and spoken text is important for synthesizing photo-realistic talking heads. Talking heads appear much more engaging when they exhibit realistic motion patterns.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115389980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An efficient method for recognition of human faces using higher orders Pseudo Zernike Moment Invariant","authors":"J. Haddadnia, M. Ahmadi, K. Faez","doi":"10.1109/AFGR.2002.1004175","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004175","url":null,"abstract":"This paper introduces a new method for the recognition of human faces in 2-dimensional digital images using a new localization of facial information and Pseudo Zernike Moment Invariants (PZMI) as features and a radial basis function (RBF) neural network as the classifier. In this paper the effect of two parameters in recognition rate improvement are studied. These include the order of the PZMI as well as facial candidate ratio (FCR) of images. The tests are carried out on the Olivetti Research Laboratory (ORL) database and a comparative study with two of the existing techniques are included to show the effectiveness of the proposed technique.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125125930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An adaptive fusion architecture for target tracking","authors":"G. Loy, L. Fletcher, N. Apostoloff, A. Zelinsky","doi":"10.1109/AFGR.2002.1004164","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004164","url":null,"abstract":"A vision system is demonstrated that adaptively allocates computational resources over multiple cues to robustly track a target in 3D. The system uses a particle filter to maintain multiple hypotheses of the target location. Bayesian probability theory provides the framework for sensor fusion, and resource scheduling is used to intelligently allocate the limited computational resources available across the suite of cues. The system is shown to track a person in 3D space moving in a cluttered environment.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"129 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128887699","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Face identification across different poses and illuminations with a 3D morphable model","authors":"V. Blanz, S. Romdhani, T. Vetter","doi":"10.1109/AFGR.2002.1004155","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004155","url":null,"abstract":"We present a novel approach for recognizing faces in images taken from different directions and under different illumination. The method is based on a 3D morphable face model that encodes shape and texture in terms of model parameters, and an algorithm that recovers these parameters from a single image of a face. For face identification, we use the shape and texture parameters of the model that are separated from imaging parameters, such as pose and illumination. In addition to the identity, the system provides a measure of confidence. We report experimental results for more than 4000 images from the publicly available CMU-PIE database.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134276725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An investigation into the use of partial-faces for face recognition","authors":"S. Gutta, V. Philomin, M. Trajkovic","doi":"10.1109/AFGR.2002.1004126","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004126","url":null,"abstract":"Even though numerous techniques for face recognition have been explored over the years, most research has primarily focused on identification from full frontal/profile facial images. This paper conducts a systemic study to assess the performance when using partial faces for identification. Our specific approach considers an ensemble of radial basis function (RBF) networks. A specific advantage of using an ensemble is its ability to cope with the inherent variability in the image formation and the data acquisition process. Our database consists of imagery corresponding to 150 unique subjects, totalling 3,000 facial images with /spl plusmn/5/spl deg/ rotation. Based on our experimental results, we observe that the average cross-validation performance is the same, even if only half the face image is used instead of the full-face image. Specifically, we obtain 96% when partial faces are used and 97% when full faces are used.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134049932","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Kale, Naresh P. Cuntoor, V. Krüger, A. Rajagopalan
{"title":"Gait-based recognition of humans using continuous HMMs","authors":"A. Kale, Naresh P. Cuntoor, V. Krüger, A. Rajagopalan","doi":"10.1109/AFGR.2002.1004176","DOIUrl":"https://doi.org/10.1109/AFGR.2002.1004176","url":null,"abstract":"Gait is a spatio-temporal phenomenon that typifies the motion characteristics of an individual. In this paper, we propose a view-based approach to recognize humans through gait. The width of the outer contour of the binarized silhouette of a walking person is chosen as the image feature. A set of stances or key frames that occur during the walk cycle of an individual is chosen. Euclidean distances of a given image from this stance set are computed and a lower-dimensional observation vector is generated. A continuous hidden Markov model (HMM) is trained using several such lower-dimensional vector sequences extracted from the video. This methodology serves to compactly capture structural and transitional features that are unique to an individual. The statistical nature of the HMM renders overall robustness to gait representation and recognition. The human identification performance of the proposed scheme is found to be quite good when tested in natural walking conditions.","PeriodicalId":364299,"journal":{"name":"Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2002-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114394171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}