{"title":"Automatic description of complex buildings with multiple images","authors":"Z. Kim, A. Huertas, R. Nevatia","doi":"10.1109/WACV.2000.895417","DOIUrl":"https://doi.org/10.1109/WACV.2000.895417","url":null,"abstract":"3-D building detection and description is a practical application of 3-D object description, a key task of computer vision. We present an approach to detecting and describing buildings of polygonal rooftops by using multiple, overlapping images of the scene. First, 3-D features are generated by using multiple images, and rooftop hypotheses are generated by neighborhood searches on those features. For robust generation of 3-D features, we present a probabilistic approach to address the epipolar alignment problem in line matching. Image-derived unedited elevation data is used to assist feature matching, and to generate rough cues of the presence of 3-D structures. These cues help reduce the search space significantly. Experimental results are shown on some complex buildings.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115412142","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Urban street grid description and verification","authors":"K. Price","doi":"10.1109/WACV.2000.895416","DOIUrl":"https://doi.org/10.1109/WACV.2000.895416","url":null,"abstract":"While two-dimensional maps exist for most urban areas, the descriptions may be incomplete or out of date, or of insufficient resolution for the given application and features such a roads are not described as 3-D objects. Most of the past work on road detection has concentrated on either low resolution, primarily rural roads (usually producing \"spaghetti\" roads with no notion of intersections), high resolution road following without the topological information of the intersections, or pixel classification where there is no sense of the road as an object. This paper address the problem of extracting a street grid in an urban environment while maintaining the topological information of the intersections. Starting from an initial seed intersection, which gives the size and orientation of the expected grid, this system uses a feature-based hypothesis and verify paradigm to extract a 3-D description of the street grid. The verification uses the context provided by an intersection model and by an extended street model and other available sensors.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"92 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115735866","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatic image segmentation and classification using on-line shape learning","authors":"Kyoung-Mi Lee, W. Street","doi":"10.1109/WACV.2000.895404","DOIUrl":"https://doi.org/10.1109/WACV.2000.895404","url":null,"abstract":"The detection, precise segmentation and classification of specific objects is an important task in many computer vision and image analysis problems, particularly in medical domains. Existing methods such as template matching typically require excessive computation and user interaction, particularly if the desired objects have a variety of different shapes. This paper presents a new approach that uses unsupervised learning to find a set of templates specific to the objects being outlined by the user. The templates are formed by averaging the shapes that belong to a particular cluster, and are used to guide an intelligent search through the space of possible objects. This results in decreased time and increased accuracy for repetitive segmentation problems, as system performance improves with continued use. Further, the information gained through clustering and user feedback is used to classify the objects for problems in which shape is relevant to the classification. The effectiveness of the resulting system is demonstrated on two applications: a medical diagnosis task using cytological images and a vehicle recognition task.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121921650","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Achieving accurate colour image segmentation in 2D and 3D with LVQ classifiers and partial adaptable class-specific representation","authors":"C. F. Nielsen, P. Passmore","doi":"10.1109/WACV.2000.895405","DOIUrl":"https://doi.org/10.1109/WACV.2000.895405","url":null,"abstract":"Adaptable Class-Specific Representation (ACSR) has previously been used as a solution to the problem of segmentation near edges in 2D colour images. Sampling windows of fixed shape used in many segmentation approaches cause misrepresentation of texture classes. ACSR greatly reduces this problem, based on simple templates, resulting in accurate semi-automatic segmentation. The price of accuracy in ACSR is high processing overhead. We introduce an initial segmentation step using a faster fixed-shape window sampling and Learning Vector Quantization, and apply ACSR only at edge point. Processing speed is significantly increased without compromising segmentation accuracy. ACSR segmentation is particularly interesting for medical applications where correct shape and size is important. We extend the ACSR framework to true 3D volume segmentation. 3D information is used for classification at all sampling points, producing better results than per slice pseudo-3D segmentation. Colour volumes based on the Visible Human Project are used to demonstrate the approach. We conclude that ACSR can produce accurate segmentation in colour 2D images and 3D volumes, and that partial ACSR can significantly reduce processing overhead without losing segmentation quality.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126503107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Image sequence analysis for real-time underwater cable tracking","authors":"A. Ortiz, M. Simo, G. Oliver","doi":"10.1109/WACV.2000.895427","DOIUrl":"https://doi.org/10.1109/WACV.2000.895427","url":null,"abstract":"Nowadays, the surveillance and inspection of underwater installations, such as power and telecommunication cables and pipelines, is carried out by operators that, being on the surface, drive a Remotely Operated Vehicle (ROV) with cameras mounted over it. This is a tedious and high time-consuming task, easily prone to errors mainly because of loss of attention or fatigue of the human operator. Besides, the complexity of the task is increased by the lack of quality of typical seabed images. In this study, the development of a vision system guiding an Autonomous Underwater Vehicle (AUV) able to detect and track automatically an underwater power cable laid on the seabed has been the main concern. The vision system that is proposed tracks the cable with an average success rate above 90%. The system has been tested using sequences coming from a video tape obtained in several tracking sessions of various real cables with a ROV driven from the surface.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126039141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Tracking and pose estimation for computer assisted localization in industrial environments","authors":"Xiang Zhang, Nassir Navab","doi":"10.1109/WACV.2000.895425","DOIUrl":"https://doi.org/10.1109/WACV.2000.895425","url":null,"abstract":"One of the common needs for many real-time augmented reality (AR) applications is the precise 'localization'. Currently at Siemens Corporate Research (SCR), we are developing a real-time system for industrial maintenance assistance. The user is moving within a large industrial site. In order to provide the user with additional information the system needs to locate the user in both real and virtual world. In this scenario, the user is equipped with a mobile computer. The objective is to track and locate the user using a camera attached to the mobile computer. With a set of coded visual markers pre-registered with the global coordinate system, the optical localization could be solved by marker detection, tracking and pose estimation. In this paper, we present the real-time marker detection and pose estimation algorithms used in our mobile localization application. To work in large and complicated industrial environments, our system needs to recover localization information from limited correspondences. We consider two different pose estimation algorithms: the homography based algorithm and the 3-point algorithm. In this paper, we present the results of the numerical experiments comparing these two methods. The experiments are carried out to determine the best approach for our application and to evaluate the accuracy and limitations of the algorithms.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"148 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131823426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Restoration of multiple images with motion blur in different directions","authors":"A. Rav-Acha, Shmuel Peleg","doi":"10.1109/WACV.2000.895398","DOIUrl":"https://doi.org/10.1109/WACV.2000.895398","url":null,"abstract":"Images degraded by motion blur can be restored when several blurred images are given, and the direction of motion blur in each image is different. Given two motion blurred images, best restoration is obtained when the directions of motion blur in the two images are orthogonal. Motion blur at different directions is common, for example, in the case of small hand-held digital cameras due to fast hand trembling and the light weight of the camera. Restoration examples are given on simulated data as well as on images with real motion blur.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114749165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Ghosal, Raghavendra Udupa, Sharath Pankanti, N. Ratha
{"title":"Learning partitioned least squares filters for fingerprint enhancement","authors":"S. Ghosal, Raghavendra Udupa, Sharath Pankanti, N. Ratha","doi":"10.1109/WACV.2000.895395","DOIUrl":"https://doi.org/10.1109/WACV.2000.895395","url":null,"abstract":"Fingerprint images contain varying amount of noise because of the limitations of the fingerprint acquisition process. It is often necessary to enhance such noisy fingerprint images so that the features extracted from them are reliable. We propose a novel approach to fingerprint enhancement where a set of filters are learned using the \"learn-from-example\" paradigm. An expert provides the ground truth information for ridges in a small set of representative fingerprint images. The space of local fingerprint patterns in a small neighborhood is partitioned into a set of expressive yet computationally simple classes. A filter is learnt for each partition by finding the optimal linear mapping (in least-square sense) from the input to the enhanced space. The proposed approach offers distinct performance and speed advantages for a wide variety of fingerprint images.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128388487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Global matching criterion and color segmentation based stereo","authors":"Hai Tao, H. Sawhney","doi":"10.1109/WACV.2000.895429","DOIUrl":"https://doi.org/10.1109/WACV.2000.895429","url":null,"abstract":"In this paper, we present a new analysis by synthesis computational framework for stereo vision. It is designed to achieve the following goals: (1) enforcing global visibility constraints, (2) obtaining reliable depth for depth boundaries and thin structures, (3) obtaining correct depth for textureless regions, and (4) hypothesizing correct depth for unmatched regions. The framework employs depth and visibility based rendering within a global matching criterion to compute depth in contrast with approaches that rely on local matching measures and relaxation. A color segmentation based depth representation guarantees smoothness in textureless regions. Hypothesizing depth from neighboring segments enables propagation of correct depth and produces reasonable depth values for unmatched region. A practical algorithm that integrates all these aspects is presented in this paper. Comparative experimental results are shown for real images. Results on new view rendering based on a single stereo pair are also demonstrated.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116389933","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Flame recognition in video","authors":"Walter Phillips, M. Shah, N. Lobo","doi":"10.1109/WACV.2000.895426","DOIUrl":"https://doi.org/10.1109/WACV.2000.895426","url":null,"abstract":"This paper presents an automatic system for fire detection in video sequences. There are many previous methods to detect fire, however, all except two use spectroscopy or particle sensors. The two that use visual information suffer from the inability to cope with a moving camera or a moving scene. One of these is not able to work on general data, such as movie sequences. The other is too simplistic and unrestrictive in determining what is considered fire, so that it can be used reliably only in aircraft dry bays. Our system uses color and motion information computed from video sequences to locate fire. This is done by first using an approach that is based upon creating a Gaussian-smoothed color histogram to determine the fire-colored pixels, and then using the temporal variation of pixels to determine which of these pixels are actually fire. Unlike the two previous vision-based methods for pre detection, our method is applicable to more areas because of its insensitivity to camera motion. Two specific applications not possible with previous algorithms are the recognition of fire in the presence of global camera motion or scene motion and the recognition of fire in movies for possible use in an automatic rating system. We show that our method works in a variety of conditions, and that it can automatically determine when it has insufficient information.","PeriodicalId":306720,"journal":{"name":"Proceedings Fifth IEEE Workshop on Applications of Computer Vision","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125697352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}