{"title":"Applying computer vision techniques to perform semi-automated analytical photogrammetry","authors":"David Nilosek, C. Salvaggio","doi":"10.1109/WNYIPW.2010.5649777","DOIUrl":"https://doi.org/10.1109/WNYIPW.2010.5649777","url":null,"abstract":"The purpose of this research is to show how common computer vision techniques can be implemented in such a way that it is possible to automate the process of analytical photogram-metry. This work develops a workflow that generates a sparse three-dimensional point cloud from a bundle of images using SIFT, RANSAC, and a sparse bundle adjustment along with basic photogrammetric methods. It then goes on to show how the output of the sparse reconstruction method can be used to generate denser three-dimensional point clouds that can be facetized and turned into high resolution three-dimensional models. This workflow was successfully tested on a five image dataset taken with RIT's WASP imaging sensor over the Van Lare wastewater treatment plant in Rochester, NY.","PeriodicalId":210139,"journal":{"name":"2010 Western New York Image Processing Workshop","volume":"191 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114082187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A framework for object class recognition with no visual examples","authors":"Grigorios Tsagkatakis, A. Savakis","doi":"10.1109/WNYIPW.2010.5649768","DOIUrl":"https://doi.org/10.1109/WNYIPW.2010.5649768","url":null,"abstract":"Traditional approaches in object class recognition utilize a large number of labeled visual examples in order to train classifiers to recognize the category of an object in a test image. However, the need for a large number of training data makes the scalability of this approach problematic. In this paper, we explore the recently proposed paradigm of attribute based category recognition for object category recognition without using any visual examples. This goal is achieved by introducing a textual based attribute representation of an image and using these attributes for object categorization. We propose the Sparse Representations (SRs) framework to achieve training-free and highly scalable attribute prediction. We investigate different approaches in mapping the predicted attributes to object classes using Nearest Neighbors and Support Vector Machines. Experimental results suggest that the use of the SRs framework in conjunction with an appropriate Nearest Neighbors scheme can improve prediction accuracy at a much lower computational cost.","PeriodicalId":210139,"journal":{"name":"2010 Western New York Image Processing Workshop","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124955140","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Propagating multi-class pixel labels throughout video frames","authors":"Albert Y. C. Chen, Jason J. Corso","doi":"10.1109/WNYIPW.2010.5649773","DOIUrl":"https://doi.org/10.1109/WNYIPW.2010.5649773","url":null,"abstract":"The effective propagation of pixel labels through the spatial and temporal domains is vital to many computer vision and multimedia problems, yet little attention have been paid to the temporal/video domain propagation in the past. Previous video label propagation algorithms largely avoided the use of dense optical flow estimation due to their computational costs and inaccuracies, and relied heavily on complex (and slower) appearance models. We show in this paper the limitations of pure motion and appearance based propagation methods alone, especially the fact that their performances vary on different type of videos. We propose a probabilistic framework that estimates the reliability of the sources and automatically adjusts the weights between them. Our experiments show that the “dragging effect” of pure optical-flow-based methods are effectively avoided, while the problems of pure appearance-based methods such the large intra-class variance is also effectively handled.","PeriodicalId":210139,"journal":{"name":"2010 Western New York Image Processing Workshop","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127567725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Jab to RGB multidimensional lookup tables","authors":"M. Maltz, Iris Zhao","doi":"10.1109/WNYIPW.2010.5649757","DOIUrl":"https://doi.org/10.1109/WNYIPW.2010.5649757","url":null,"abstract":"The CIECAM color appearance space (Jab) is widely used in color science. The relationship between XYZ and Jab is defined by a set of equations. For ease of computation and architectural flexibility, it is desirable to implement these relationships in multi-dimensional lookup table form. Unfortunately, for some colors just outside the sRgb gamut, these equations give complex numbers. Furthermore, to include colors that printers can produce, a transformation between Jab and a wider gamut RGB is necessary. Such a table would contain problematical colors. To solve this problem we accurately reproduce the CIECAM results within a gamut of reasonable colors. Nodes just outside the gamut are used when interpolating to find colors just inside the gamut so they are also affected by the accuracy constraint. All the nodes outside the gamut are also subjected to a smoothness constraint. Both RGB to Jab and Jab to RGB tables have been produced for a wide gamut RGB, and give quite accurate results.","PeriodicalId":210139,"journal":{"name":"2010 Western New York Image Processing Workshop","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133227426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Anchor point selection by KL-divergence","authors":"Charles Parker","doi":"10.1109/WNYIPW.2010.5649752","DOIUrl":"https://doi.org/10.1109/WNYIPW.2010.5649752","url":null,"abstract":"Selecting anchor points for the identification of scanned documents can be an effective and quick means of identifying unknown documents. Here, we discuss and compare some strategies for classification of scanned forms using anchor points and show experiments indicating that a robust system can be built with only a few training examples.","PeriodicalId":210139,"journal":{"name":"2010 Western New York Image Processing Workshop","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128112036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Calibration of a multi-projector system for display on a cylindrical surface","authors":"Brandon B. May, N. Cahill, M. Rosen","doi":"10.1109/WNYIPW.2010.5649778","DOIUrl":"https://doi.org/10.1109/WNYIPW.2010.5649778","url":null,"abstract":"In this paper we present a method for geometrically, photometrically & colorimetrically calibrating a multi-projector system for display on a cylindrical surface. Using a single camera we reconstruct the 3D surface of the display and determine the projector-screen relationship to accurately register projected images. Given these relationships, we apply chrominance gamut morphing in the overlap regions to smoothly transition from one projected image to the next. After white point balancing and perceptual brightness constraining, the final registered and blended images are shown by each respective projector to create a seamless high resolution image.","PeriodicalId":210139,"journal":{"name":"2010 Western New York Image Processing Workshop","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115344168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Utilizing the graph modularity to blind cluster multispectral satellite imagery","authors":"Ryan A. Mercovich, A. Harkin, D. Messinger","doi":"10.1109/WNYIPW.2010.5649737","DOIUrl":"https://doi.org/10.1109/WNYIPW.2010.5649737","url":null,"abstract":"The fully automatic separation of spectral image data into clusters is a problem with a wide variety of desired and potential solutions. Contrary to the typical approaches of utilizing the first order statistics and Gaussian modeling of spectral image data, the method described in this paper utilizes the spectral data structure to generate a graph representation of the image and then clusters the data by applying the method of optimal modularity for finding communities within the graph. After defining and identifying pixel adjacencies to represent an image as an adjacency matrix, a quantity known as the graph modularity is maximized to split the data into groups of spectrally similar pixels. Recursion with the subgroups created in each split creates the data clustering. The groups where the maximal modularity is not above a given threshold are not split, resulting in a stopping condition and an estimation of the number of clusters necessary to fully describe the data. By ignoring any reliance on assumptions of the shape of the data, this method excels in regions where typical clustering methods fail, such as high resolution urban scenes with very high clutter and regions with subtle variability like coastal bodies of water.","PeriodicalId":210139,"journal":{"name":"2010 Western New York Image Processing Workshop","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114447900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A euclidean distance transformation for improved anomaly detection in spectral imagery","authors":"A. Schlamm, D. Messinger","doi":"10.1109/WNYIPW.2010.5649762","DOIUrl":"https://doi.org/10.1109/WNYIPW.2010.5649762","url":null,"abstract":"Remotely sensed spectral imagery is used in many disciplines, including environmental monitoring, agricultural health, defense and security applications, astronomy, medical imaging, and food quality assessment. The basic tasks performed within any of these fields are target or anomaly detection, classification or clustering, change detection, and physical parameter estimation. Hyperspectral image (HSI) analysis in any field often involves mathematically transforming the raw data into a new space using Principal Components Analysis (PCA) or similar techniques. The dimensionality of this new space is usually smaller than the collected space and as a result reduces computation time of subsequent algorithms. Additionally and more importantly, the results of standard algorithms may perform better in this new, uncor-related space. Many of the currently used transformations in HSI analysis are statistical in nature and therefore place Gaussian or similar assumptions on the data distribution. These assumptions work well with remote sensing imagery with low spectral and/or spatial resolution. In low resolution imagery, each pixel is a mixture of many materials and the data distribution is often sufficiently represented by statistical distributions. However, as the current generation sensors typically have higher spatial and/or spectral resolution, the complexity of the data collected is increasing and these assumptions are no longer adequate. As a result, algorithms based on these statistical transformations do not necessarily provide improved results on modern datasets. A new, data driven, mathematical transformation is presented as a preprocessing step for HSI analysis. Termed the Nearest Neighbor Transformation, this new transformation does not rely on placing assumptions on the data and may improve analytical results from standard HSI anomaly detection algorithms.","PeriodicalId":210139,"journal":{"name":"2010 Western New York Image Processing Workshop","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131719884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}