{"title":"Face detection and attentional frames for visually mediated interaction","authors":"A. Howell, H. Buxton","doi":"10.1109/HUMO.2000.897384","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897384","url":null,"abstract":"Introduces a set of effective computational techniques for understanding the visual aspects of human interaction, which could be used, for example, in videoconferencing applications. First, we present methods for face detection and the capture of attentional frames to focus the processing for visually mediated interaction. Second, we present methods for recognising the various gesture phases that can be used to control the camera systems in the integrated system. Finally, we discuss how these techniques can be extended to \"virtual groups\" of multiple people interacting at multiple sites.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124477451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A computational model for motion detection and direction discrimination in humans","authors":"Yang Song, P. Perona","doi":"10.1109/HUMO.2000.897364","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897364","url":null,"abstract":"Seeing biological motion is very important for both humans and computers. Psychophysics experiments show that the ability of our visual system for biological motion detection and direction discrimination is different from that for simple translation. The existing quantitative models of motion perception cannot explain these findings. We propose a computational model, which uses learning and statistical inference based on the joint probability density function (PDF) of the position and motion of the body, on stimuli similar to (Neri et al., 1998). Our results are consistent with the psychophysics indicating that our model is consistent with human motion perception, accounting for both biological motion and pure translation.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"362 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126698069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Camera handoff: tracking in multiple uncalibrated stationary cameras","authors":"O. Javed, Sohaib Khan, Z. Rasheed, M. Shah","doi":"10.1109/HUMO.2000.897380","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897380","url":null,"abstract":"Multiple cameras are needed to completely cover an environment for monitoring activity. To track people successfully in multiple perspective imagery, one needs to establish a correspondence between objects captured by multiple cameras. We present a system for tracking people using multiple uncalibrated cameras. The system is able to discover spatial relationships between the cameras' fields of view and to use this information to correspond between different perspective views of the same person. We employ the novel approach of finding the limits of the field of view of a camera that are visible by the other cameras. This helps us to disambiguate between possible candidates of correspondence. The proposed approach is very fast compared to camera calibration-based approaches.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130676756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Real-time human motion analysis and IK-based human figure control","authors":"S. Yonemoto, Daisaku Arita, R. Taniguchi","doi":"10.1109/HUMO.2000.897385","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897385","url":null,"abstract":"The paper presents real-time human motion analysis based on real-time inverse kinematics. Our purpose is to realize a mechanism of human-machine interaction via human gestures, and, as a first step, we have developed a computer-vision-based human motion analysis system. In general, man-machine \"smart\" interaction requires a real-time human full-body motion capturing system without special devices or markers. However, since such a vision-based human motion capturing system is essentially unstable and can only acquire partial information because of self-occlusion, we have to introduce a robust pose estimation strategy, or an appropriate human motion synthesis based on motion filtering. To solve this problem, we have developed a method based on inverse kinematics, which can estimate human postures with limited perceptual cues such as positions of a head, hands and feet. We outline a real-time and on-line human motion capture system and demonstrate a simple interaction system based on the motion capture system.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"122 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132406848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hand shape estimation using image transition network","authors":"Y. Hamada, N. Shimada, Y. Shirai","doi":"10.1109/HUMO.2000.897387","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897387","url":null,"abstract":"We present a method of hand posture estimation from silhouette images taken by two cameras. First, we extract the silhouette contour for a pair of images. We construct an eigenspace from images of hands with various postures. For effective matching, we define a shape complexity for each image to see how well the shape feature is represented. For a pair of input images, the total matching error is computed by combining the two matching errors according to the shape complexity. Thus the best-matched image is obtained for a pair of images. For rapid processing, we limit the matching candidate by using the constraint on the shape change. The possible shape transition is represented by a transition network. Because the network is hard to build, we apply offline learning, where nodes and links are automatically created by showing examples of hand shape sequences. We show experiments of building the transition networks and the performance of matching using the network.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133665577","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A framework for motion recognition with applications to American sign language and gait recognition","authors":"Christian Vogler, H. C. Sun, Dimitris N. Metaxas","doi":"10.1109/HUMO.2000.897368","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897368","url":null,"abstract":"Human motion recognition has many important applications, such as improved human-computer interaction and surveillance. A big problem that plagues this research area is that human movements can be very complex. Managing this complexity is difficult. We turn to American sign language (ASL) recognition to identify general methods that reduce the complexity of human motion recognition. We present a framework for continuous 3D ASL recognition based on linguistic principles, especially the phonology of ASL. This framework is based on parallel hidden Markov models (HMMs), which are able to capture both the sequential and the simultaneous aspects of the language. Each HMM is based on a single phoneme of ASL. Because the phonemes are limited in number, as opposed to the virtually unlimited number of signs that can be composed from them, we expect this framework to scale well to larger applications. We then demonstrate the general applicability of this framework to other human motion recognition tasks by extending it to gait recognition.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123219554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"ELEVIEW: an active elevator video surveillance system","authors":"Hui Shao, Liyuan Li, Ping Xiao, M. Leung","doi":"10.1109/HUMO.2000.897373","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897373","url":null,"abstract":"In this paper, a novel study for an automated scene interpretation system, named ELEVIEW, is reported to outline the design of the system. It is motivated by the reported crimes that happen inside elevators. The main goal is to investigate techniques that make an ordinary elevator monitoring system intelligent, i.e. see the scene and understand actions that are occurring. The system could filter out normal actions and trigger an alarm once abnormal events are detected. The paper focuses on the system overview, segmentation techniques as well as scenario classification. A double thresholded segmentation is employed to enhance the segmentation outcomes. This paper mainly presents an overview of the system and significant results so far achieved.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"05 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130865229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Tracking multiple objects in the presence of articulated and occluded motion","authors":"S. Dockstader, Murat Tekalp","doi":"10.1109/HUMO.2000.897376","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897376","url":null,"abstract":"Presents a novel approach to the tracking of multiple articulate objects in the presence of occlusion in moderately complex scenes. Most conventional tracking algorithms work well when only one object is tracked at a time. However, when multiple objects must be tracked simultaneously, significant computation is often introduced in order to handle occlusion and to calculate the appropriate region correspondence between successive frames. We introduce a near-real-time solution to this problem by using a probabilistic mixing of low-level features and components. The algorithm mixes coarse motion estimates, change detection information and unobservable predictions to create accurate trajectories of moving objects. We implement this multifeature mixing strategy within the context of a video surveillance system using a modified Kalman filtering mechanism. Experimental results demonstrate the efficacy of the proposed tracking and surveillance system.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134055289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
J. Wilhelms, A. V. Gelder, L. Atkinson-Derman, A. Luo
{"title":"Human motion from active contours","authors":"J. Wilhelms, A. V. Gelder, L. Atkinson-Derman, A. Luo","doi":"10.1109/HUMO.2000.897386","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897386","url":null,"abstract":"We describe an approach for extracting three-dimensional articulated motion from unrestricted monocular video sequences. We combine feature extraction methods based on active contours with interactive adjustment. An articulated model is interactively aligned with the image in selected anchor frames. Active contour points are anchored to model segments in these frames. Occluded points are detected using object geometry and do not participate in edge tracking. Model joints are automatically adjusted in other frames to align with active contour points. The combination of interactive and automatic adjustment allows extraction of arbitrarily complex movements.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"234-235 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123977732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Human activity detection in MPEG sequences","authors":"B. Ozer, W. Wolf, A. Akansu","doi":"10.1109/HUMO.2000.897372","DOIUrl":"https://doi.org/10.1109/HUMO.2000.897372","url":null,"abstract":"We propose a hierarchical method for human detection and activity recognition in MPEG sequences. The algorithm consists of three stages at different resolution levels. The first step is based on the principal component analysis of MPEG motion vectors of macroblocks grouped according to velocity, distance and human body proportions. This step reduces the complexity and amount of processing data. The DC DCT components of luminance and chrominance are the input for the second step, to be matched to activity templates and a human skin template. A more detailed analysis of the uncompressed regions extracted in previous steps is done at the last step via model-based segmentation and graph matching. This hierarchical scheme enables working at different levels, from low complexity to low false rates. It is important and interesting to realize that significant information can be obtained from the compressed domain in order to connect to high level semantics.","PeriodicalId":384462,"journal":{"name":"Proceedings Workshop on Human Motion","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128925926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}