{"title":"Relative depth from motion using normal flow: an active and purposive solution","authors":"Liuqing Huang, Y. Aloimonos","doi":"10.1109/WVM.1991.212807","DOIUrl":"https://doi.org/10.1109/WVM.1991.212807","url":null,"abstract":"The authors show how an active observer can compute the relative depth of (stationary or moving) objects in the field of view using only the spatiotemporal derivatives of the time varying image intensity function. The solution they propose is purposive in the sense that it solves only the relative depth from motion problem and cannot be used for other problems related to motion; active in the sense that the activity of the observer is essential for the solution of the problem. Results indicate that exact computation of retinal motion does not appear to be a necessary first step for some problems related to visual motion. In addition, optic flow, whose computation is an ill-posed problem, is related to the motion of the scene only under very restrictive assumptions. As a result, the use of optic flow in some quantitative motion analysis studies is questionable.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128844633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Non-rigid motion and structure from contour","authors":"Alex Pentland, Bradley Horowitz, Stan Sclaroff","doi":"10.1109/WVM.1991.212773","DOIUrl":"https://doi.org/10.1109/WVM.1991.212773","url":null,"abstract":"The elastic properties of real materials constrain the types of non-rigid motion that can occur. The authors show that by modeling and simulating the physics of non-rigid motion they can obtain good estimates of both object shape and velocity from contour data. The underlying model is based on the finite element method, and decouples the degrees of freedom by breaking down object motion into rigid and non-rigid vibration modes. Shape estimates are integrated over time by use of an extended Kalman filter, resulting in a stable estimate of both 3D shape and 3D velocity. Examples using X-ray imagery (of the heat) are presented.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"73 4 Pt 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131818577","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D structure reconstruction from an ego motion sequence using statistical estimation and detection theory","authors":"Y.-L. Chang, J. Aggarwal","doi":"10.1109/WVM.1991.212797","DOIUrl":"https://doi.org/10.1109/WVM.1991.212797","url":null,"abstract":"The paper discusses the problem of estimating 3D structures from an extended sequence of 2D images taken by a moving camera with known motion. The work is mainly concerned with sparse line features and thus is a natural extension of the feature-based motion analysis paradigm. Usually such a paradigm involves several separate operations: feature detection, feature matching, structure/motion estimation, and higher level processing, such as feature grouping. The authors propose to integrate the different phases based on the statistical estimation and detection theory. They show how each operation can be formalized and, in particular, consider the structure parameter estimation and the feature matching together as the combined estimation-decision problem. The proposed algorithm is tested with both synthetic and real data.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115779481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Forming a three dimensional environment model using multiple observations","authors":"P. Khalili, R. Jain","doi":"10.1109/WVM.1991.212798","DOIUrl":"https://doi.org/10.1109/WVM.1991.212798","url":null,"abstract":"An autonomous navigating agent must form a three-dimensional model of its environment using passive sensors. Typical stereo algorithms produce sparse depth maps and cannot be used to distinguish between holes and solid objects in the environment. The authors present a novel methodology for creating a three-dimensional model of the environment. They divide the environment into a set of disjoint cells. Using multiple images obtained from different view points, they estimate the mean and variance of intensity observed for each cell. The computed variance can be used to distinguish between empty and full cells in the environment. The technique, unlike the typical stereo methodology, does not rely on solving the correspondence problem. The resulting model of the environment is dense and can be used directly for navigation. Experimental results are presented.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"107 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125048097","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Mechanisms for isolating component patterns in the sequential analysis of multiple motion","authors":"Peter J. Burt, Rajesh Hingorani, R. Kolczynski","doi":"10.1109/WVM.1991.212808","DOIUrl":"https://doi.org/10.1109/WVM.1991.212808","url":null,"abstract":"Pyramid techniques are commonly used to provide computational efficiency in the analysis of image motion. But these techniques can play an even more important role in the analysis of multiple motion, where, for example, a transparent pattern moves in front of a differently moving background pattern. The pyramid framework then separates motion components based on their spatial and temporal frequency characteristics so that each can be estimated independently of the others. This property is key to recently proposed selective stabilization algorithms for the sequential analysis of multiple motion and for the detection of moving objects from a moving platform. The authors determine the conditions for component selection. Results can provide important guidance in practical applications of motion analysis.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130560956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hierarchical feature based matching for motion correspondence","authors":"V. Venkateswar, R. Chellappa","doi":"10.1109/WVM.1991.212774","DOIUrl":"https://doi.org/10.1109/WVM.1991.212774","url":null,"abstract":"The authors design a feature based motion correspondence system. They propose a hierarchical grouping process that groups line segments into more complex structures that are easier to match. The hierarchy consists of lines, vertices, edges and surfaces. Matching starts at the highest level of the hierarchy (surfaces) and proceeds to the lowest (lines). Higher level features are easier to match, because they are fewer in number and more distinct in form. These matches then constrain the matches at lower levels. Perceptual and structural relations are used to group matches into islands of certainty. A Truth Maintenance System (TMS) is used to enforce grouping constraints and eliminates inconsistent match groupings. The TMS is also used for reasoning in the presence of uncertainty and to carry out logic revisions necessitated by additions, deletions and confirmations of hypotheses. The hierarchical matching process results in line matches as well as point matches. These then can be used as an input to a motion estimation algorithm.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132614730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Parallel visual motion analysis using multiscale Markov random fields","authors":"Fabrice Heitz, Patrick Perez, P. Bouthemy","doi":"10.1109/WVM.1991.212791","DOIUrl":"https://doi.org/10.1109/WVM.1991.212791","url":null,"abstract":"The use of Markov Random Field (MRF) models within the framework of global bayesian decision has brought new powerful solutions to visual motion analysis. The efficiency of MRF models for image sequence analysis has been proved on various classes of real-world sequences: outdoor and indoor scenes including several moving objects and camera motion. The authors extend this work by investigating new multiscale motion analysis algorithms based on MRF models. These algorithms are related to a new class of consistent multiscale MRF statistical models. The multiscale paradigm exhibits fast convergence properties towards quasi optimal estimates. Its performances are compared to standard relaxation in the case of optical flow measurement.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134236544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Estimating 3D vehicle motion in an outdoor scene from monocular and stereo image sequences","authors":"M. K. Leung, Yuncai Liu, T. S. Huang","doi":"10.1109/WVM.1991.212787","DOIUrl":"https://doi.org/10.1109/WVM.1991.212787","url":null,"abstract":"The main goal of this research is to test how well existing feature extraction, matching and motion estimation algorithms (with appropriate modification) work on outdoor scenes. For this purpose, a careful calibrated image sequence data base has been created. The data used for the results reported in the paper consists of a sequence of 24 stereo images of an outdoor scene containing a moving truck with stationary background. Two motion estimation methods using feature correspondences were applied in the data: point correspondences over two stereo image pairs and line correspondences over three monocular images. In spite of the large values of the range to baseline ration (10:1) and the range to focal length ration (1000:1), the estimated rotation parameters are reasonably accurate (10-20% errors) in both methods. Although the translation estimates in the monocular method are large, the translation errors in the stereo method are around 1 meter, and are mainly due to image sampling.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123355061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Direct multi-resolution estimation of ego-motion and structure from motion","authors":"K., J. Hanna, David Sarnoff","doi":"10.1109/WVM.1991.212812","DOIUrl":"https://doi.org/10.1109/WVM.1991.212812","url":null,"abstract":"The paper describes an iterative algorithm that estimates the motion of a camera through an environment directly from brightness derivatives of an image pair. A global ego-motion constraint is combined with the local brightness constancy constraint to relate local surface models with the global ego-motion model and local brightness derivatives. In an iterative process, the author first refines the local surface models using the ego-motion as a constraint, and then refines the ego-motion model using the local surface models as constraints. He performs this analysis at multiple resolutions. He shows how information from local corner-like and edge-like image structures contribute to the refinement of the global ego-motion estimate, and how the ego-motion constraint can help resolve local motion ambiguities that arise from the aperture problem. Results of the algorithm are shown on uncalibrated outdoor image sequences, and also on a computer-rendered image sequence.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125495418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust estimation of a multi-layered motion representation","authors":"Trevor Darrell, A. Pentland","doi":"10.1109/WVM.1991.212810","DOIUrl":"https://doi.org/10.1109/WVM.1991.212810","url":null,"abstract":"In order to recover an accurate representation of a scene containing multiple moving objects, one must use estimation methods that can recover both model parameters and segmentation at the same time. Traditional approaches to this problem rely on an edge-based discontinuity model, and have problems with transparent phenomena. The authors introduce a layered model of scene segmentation based on explicitly representing the support of a homogeneous region. The model employs parallel robust estimation techniques, and uses a minimal-covering optimization to estimate the number of objects in the scene. Using a simple direct motion model of translating objects, they successfully segment real image sequences containing multiple motions.<<ETX>>","PeriodicalId":208481,"journal":{"name":"Proceedings of the IEEE Workshop on Visual Motion","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1991-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122166995","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}