{"title":"Real-time visually guided human figure control using IK-based motion synthesis","authors":"S. Yonemoto, Daisaku Arita, R. Taniguchi","doi":"10.1109/WACV.2000.895422","DOIUrl":"https://doi.org/10.1109/WACV.2000.895422","url":null,"abstract":"This paper presents a real-time human figure motion control method using color blob tracking, and human motion synthesis based on real-time inverse kinematics. Our purpose is to do seamless mapping of human motion in the real world into virtual environments. In general, virtual environment applications such as man-machine 'smart' interaction require real-time human full-body motion capturing systems without special devices or markers. However, since such vision-based human motion capturing systems are essentially unstable and can only acquire partial information because of self-occlusion, we have to introduce a robust pose estimation strategy, or an appropriate human motion synthesis based on motion filtering. In this paper, we have demonstrated a real-time and online real-virtual interaction system which realizes human full-body motion capturing and synthesis.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114940106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Modeling 3-D complex buildings with user assistance","authors":"S. Lee, A. Huertas, R. Nevatia","doi":"10.1109/WACV.2000.895419","DOIUrl":"https://doi.org/10.1109/WACV.2000.895419","url":null,"abstract":"An effective 3D method incorporating user assistance for modeling complex buildings is proposed. This method utilizes the connectivity and similar structure information among unit blocks in a multi-component building structure, to enable the user to incrementally construct models of many types of buildings. The system attempts to minimize the time and the number of user interactions needed to assist an existing automatic system in this task. Several examples are presented that demonstrate significant improvement and efficiency compared with other approaches and with purely manual systems.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121854000","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An agent network for microburst detection","authors":"S. Dance","doi":"10.1109/WACV.2000.895409","DOIUrl":"https://doi.org/10.1109/WACV.2000.895409","url":null,"abstract":"Meteorological data provides a rich source of imagery which is well suited to the growing field of agent networks based image processing. We discuss one such system which analyses Doppler and reflectivity radar volumes to detect microbursts (windshear events). The system employs a network of agents, each performing a subtask of the detection, with message passing between agents. The various agents detect reflectivity cores, high divergent shear zones, microbursts and trade microbursts through time. It is implemented in the Java language, and is successfully detecting microbursts in data from the Kurnell radar near Sydney.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125010416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Creating 3D models with uncalibrated cameras","authors":"Mei Han, T. Kanade","doi":"10.1109/WACV.2000.895420","DOIUrl":"https://doi.org/10.1109/WACV.2000.895420","url":null,"abstract":"We describe a factorization-based method to recover 3D models from multiple perspective views with uncalibrated cameras. The method first performs a projective reconstruction using a bilinear factorization algorithm, and then converts the projective solution to a Euclidean one by enforcing metric constraints. We present three factorization-based normalization algorithms to generate the Euclidean reconstruction and the intrinsic parameters, assuming zero skews. The first two algorithms are linear, one for dealing with the case that only the focal lengths are unknown, and another for the case that the focal lengths and the constant principal point are unknown. The third algorithm is bilinear dealing with the case that the focal lengths, the principal points and the aspect ratios are all unknown. We present the results of applying this method to building modeling, terrain recovery and multi-camera calibration.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132394278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A robust background subtraction method for changing background","authors":"M. Seki, H. Fujiwara, K. Sumi","doi":"10.1109/WACV.2000.895424","DOIUrl":"https://doi.org/10.1109/WACV.2000.895424","url":null,"abstract":"Background subtraction is a useful and effective method for detecting moving objects in video images. Since this method assumes that image variations are caused only by moving objects (i.e., the background scene is assumed to be stationary), however, its applicability is limited. In this paper, we propose a background subtraction method that robustly handles various changes in the background. The method learns the chronological changes in the observed scene's background in terms of distributions of image vectors. The method operates the subtraction by evaluating the Mahalanobis distances between the averages of such image vectors and newly observed image vectors. The method we propose herein expresses actual changes in the background using a multi-dimensional image vector space. This enables the method to detect objects with the correct sensitivity. We also introduce an eigenspace to reduce the computational cost. We describe herein how approximate Mahalanobis distances are obtained in this eigenspace. In our experiments, we confirmed the proposed method's effectiveness for real world scenes.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130242051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Synthesized virtual view-based eigenspace for face recognition","authors":"Jie Yan, HongJiang Zhang","doi":"10.1109/WACV.2000.895407","DOIUrl":"https://doi.org/10.1109/WACV.2000.895407","url":null,"abstract":"This paper presents a new face recognition method using virtual view-based eigenspace. This method provides a possible way to recognize human face of different views even when samples of a view are not available. To achieve this, we have developed a virtual human face generation technique that synthesizes human face of arbitrary views. By using a frontal and profile images of a specific subject, a deformation technique allows automatic alignment of features in the 3-D generic graphic face model with the features of the pre-provided images of the specific subject. The deformation result is a 3-D face model of the specific human face. It reflects accurately the correspondence geometric features and texture features of the specific subject. In the recognition step, we use an extended nearest-neighbor rule based on an Euclidean distance measure as the recognition classifier. This work shows the feasibility of applying 3-D modeling techniques onto face recognition problems.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"164 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114150022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A flexible eyetracker for psychological applications","authors":"D. DeVault, A. H. Bond","doi":"10.1109/WACV.2000.895423","DOIUrl":"https://doi.org/10.1109/WACV.2000.895423","url":null,"abstract":"We describe a practical method for measuring eye movements during psychological tests. This is an important class of applications including clinical evaluations and marketing studies. Existing methods in common use for psychological measurement, for example infrared reflection methods, are invasive involving head stabilization and special purpose lighting. In our experiments, we need to observe subjects for long periods, on the order of one hour. In addition, subjects must verbalize, which makes it difficult to stabilize their heads relative to the camera. We track the head using a lightweight spectacle framework worn by the subject. It has a set of easily visible colored balls. We segment each image into four characteristic colors, corresponding to iris, yellow ball, red ball, and background, which are obtained by sampling the images for each subject. The classification into colors is done by training a simple neural network for each characteristic color. We match a template to color-reduced image regions to find the balls and the two irises. We use a model-based object pose method, which uses a prior measurement of the relative positions of the balls, to calculate the spectacle framework pose (the head pose). A linear method is used for calibrating gaze position against head pose and iris positions. The subject's gaze position can be traded reliably for periods of more than an hour. The locations of image features are found with an accuracy of approximately one pixel of the image. In a 640/spl times/480 image of the whole face, the eyes are each about 80 pixels across. This gives a corresponding accuracy of calculated eye gaze position on a 17 inch monitor of about 1 cm horizontally and 2 cm vertically. This method has shown itself in practice to be very flexible for psychological measurement, giving sufficient accuracy and being noninvasive.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127550797","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Detection of side-view faces in color images","authors":"Gang Wei, Dongge Li, I. Sethi","doi":"10.1109/WACV.2000.895406","DOIUrl":"https://doi.org/10.1109/WACV.2000.895406","url":null,"abstract":"A coarse-to-fine scheme for the detection of sideview faces in color images is proposed in this paper, which extends the current state of the art of face detection research. The input image can be of complex scene, containing cluttered background and confusing objects. The system consists of four stages, each of which is a refinement of the previous one, namely: (1) skin-tone detection by color, (2) region and edge preprocessing with morphological operations and length filtering, (3) face candidate region selection based on normalized similarity value and (4) final verification using hidden Markov models. Encouraging experimental results have been obtained, due to the utilization of multiple features of the input image and the conjunction of employment of various image processing and pattern recognition techniques. Besides providing the ability to detect faces other than frontal-view, our work has 3 original contributions, including the Normalized Similarity Value (NSV) to detect the presence of a given curve pattern, the iterative partition process to segment the object from confusing extraneous regions for higher detection accuracy and the exploration of the use of HMM to recognize objects in images.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"115 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116969985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multimodal image registration using local frequency","authors":"Jonathan Liu, B. Vemuri, F. Bova","doi":"10.1109/WACV.2000.895412","DOIUrl":"https://doi.org/10.1109/WACV.2000.895412","url":null,"abstract":"Fusing of multi-modal data involves automatically estimating the coordinate transformation required to align the multi-modal image data sets. Most existing methods in literature are not fast enough (take hours for estimating nonrigid deformations) for practical use. We propose a very fast algorithm, based on matching local-frequency image representations, which naturally allows for processing the data at different scales/resolutions, a very desirable property from a computational efficiency view point. This algorithm involves minimizing-over all affine transformations-the expectation of the squared difference between the local-frequency representations of the source and target images. In cases where fusing the multi-modal data requires estimating the non-rigid deformations, we propose a novel and fast PDE-based morphing technique that will estimate this non-rigid alignment. We present implementation results for synthesized and real misalignments between CT and MR brain scans. In both the cases, we validate our results against ground truth registrations which for the former case are known and for the latter are obtained from manual registration performed by an expert.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123693074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Assessing the authorship confidence of handwritten items","authors":"Sung-Hyuk Cha, S. Srihari","doi":"10.1109/WACV.2000.895401","DOIUrl":"https://doi.org/10.1109/WACV.2000.895401","url":null,"abstract":"The Writer Verification is a process to compare questioned handwriting with samples of handwriting obtained from known sources for the purposes of determining authorship or non-authorship. It plays an important investigative and forensic role in many types of crime. Particularly problematic in this regard is the fact that there exists neither true standard nor universal definition of comparison. For this reason, we propose an algorithmic objective approach. Using visual information of two or more digitally scanned handwritten items, we show a method to access the authorship confidence that is the probability of errors. Instead of building a costly handwritten item database to support the confidence, questioned words are simulated from the CEDAR letter image database in order to handle any handwritten items. An Artificial Neural Network is trained to verify the authorship using the synthesized words.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"258 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116107780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}