{"title":"Detecting plains and Grevy's Zebras in the realworld","authors":"J. Parham, C. Stewart","doi":"10.1109/WACVW.2016.7470122","DOIUrl":"https://doi.org/10.1109/WACVW.2016.7470122","url":null,"abstract":"Photographic censusing can be partly automated by leveraging the power of computer vision detection algorithms. Detecting zebras in the real world can be challenging due to varying viewpoints of the animal, natural and artificial occlusions, and overlapping animals. To address these challenges, we evaluate three detection algorithms: Hough Forests by [8], the YOLO network by [20], and Faster R-CNN [21]. We train the detectors on a soon-to-be-released dataset of 2,500 images containing 3,541 bounding boxes of plains zebras (Equus quagga) and 2,672 bounding boxes of Grevy's zebras (Equus grevyi). The detection errors are analyzed by species, viewpoint, and density (the number of bounding boxes per image). The best detector in our evaluation reports a detection mAP of 55.6% for plains and 56.6% for Grevy's.","PeriodicalId":185674,"journal":{"name":"2016 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"318 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123234528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
David C. Zhang, Giorgos Kopanas, C. Desai, S. Chai, M. Piacentino
{"title":"Unsupervised underwater fish detection fusing flow and objectiveness","authors":"David C. Zhang, Giorgos Kopanas, C. Desai, S. Chai, M. Piacentino","doi":"10.1109/WACVW.2016.7470121","DOIUrl":"https://doi.org/10.1109/WACVW.2016.7470121","url":null,"abstract":"Scientists today face an onerous task to manually annotate vast amount of underwater video data for fish stock assessment. In this paper, we propose a robust and unsupervised deep learning algorithm to automatically detect fish and thereby easing the burden of manual annotation. The algorithm automates fish sampling in the training stage by fusion of optical flow segments and objective proposals. We auto-generate large amounts of fish samples from the detection of flow motion and based on the flow-objectiveness overlap probability we annotate the true-false samples. We also adapt a biased training weight towards negative samples to reduce noise. In detection, in addition to fused regions, we used a Modified Non-Maximum Suppression (MNMS) algorithm to reduce false classifications on part of the fishes from the aggressive NMS approach. We exhaustively tested our algorithms using NOAA provided, luminance-only underwater fish videos. Our tests have shown that Average Precision (AP) of detection improved by about 10% compared to non-fusion approach and about another 10% by using MNMS.","PeriodicalId":185674,"journal":{"name":"2016 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131225561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Low resolution vehicle re-identification based on appearance features for wide area motion imagery","authors":"Mickael Cormier, L. Sommer, Michael Teutsch","doi":"10.1109/WACVW.2016.7470114","DOIUrl":"https://doi.org/10.1109/WACVW.2016.7470114","url":null,"abstract":"The description of vehicle appearance in Wide Area Motion Imagery (WAMI) data is challenging due to low resolution and renunciation of color. However, appearance information can effectively support multiple object tracking or queries in a real-time vehicle database. In this paper, we present a systematic evaluation of existing appearance descriptors that are applicable to low resolution vehicle reidentification in WAMI data. The problem is formulated as a one-to-many re-identification problem in a closed-set, where a query vehicle has to be found in a list of candidates that is ranked w.r.t. their matching similarity. For our evaluation we use a subset of the WPAFB 2009 dataset. Most promising results are achieved by a combined descriptor of Local Binary Patterns (LBP) and Local Variance Measure (VAR) applied to local grid cells of the image. Our results can be used to improve appearance based multiple object tracking algorithms and real-time vehicle database search algorithms.","PeriodicalId":185674,"journal":{"name":"2016 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131667313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Image Surveillance Assistant","authors":"Michael Maynord, S. Bhattacharya, D. Aha","doi":"10.1109/WACVW.2016.7470119","DOIUrl":"https://doi.org/10.1109/WACVW.2016.7470119","url":null,"abstract":"Security watchstanders who monitor multiple videos over long periods of time can be susceptible to information overload and fatigue. To address this, we present a configurable perception pipeline architecture, called the Image Surveillance Assistant (ISA), for assisting watchstanders with video surveillance tasks. We also present ISA1, an initial implementation that can be configured with a set of context specifications which watchstanders can select or provide to indicate what imagery should generate notifications. ISA1's inputs include (1) an image and (2) context specifications, which contain English sentences and a decision boundary defined over object detection vectors. ISA1 assesses the match of the image with the contexts by comparing (1) detected versus specified objects and (2) automatically-generated versus specified captions. Finally, we present a study to assess the utility of using captions in ISA1, and found that they substantially improve the performance of image context detection.","PeriodicalId":185674,"journal":{"name":"2016 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134609464","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Scene-independent feature- and classifier-based vehicle headlight and shadow removal in video sequences","authors":"Qun Li, Edgar A. Bernal, Matthew Shreve, R. Loce","doi":"10.1109/WACVW.2016.7470115","DOIUrl":"https://doi.org/10.1109/WACVW.2016.7470115","url":null,"abstract":"Detection of moving and foreground objects is a key step in video-based object tracking within computer vision applications such as surveillance and traffic monitoring. Foreground object detection and segmentation is usually performed based on appearance. Hence, significant detection errors can be incurred due to shadows and light sources. Most existing shadow detection algorithms exploit a large set of assumptions to limit complexity, and at the same time, rely on carefully selected parameters either in the shadow model or the decision threshold. This limits their accuracy and extensibility to different scenarios. Furthermore, most traditional shadow detection algorithms operate on each pixel in the originally detected foreground mask and make pixel-wise decisions, which is not only time-consuming but also error-prone. Little work has been done to address false foreground detection caused by vehicle headlights during nighttime. In this paper, we introduce an efficient and effective algorithm for headlight/shadow removal in modelbased foreground detection via background estimation and subtraction. The underlying assumption is that headlights and shadows do not significantly affect the texture of the background. We train a classifier to discriminate between background affected and unaffected by shadows or headlights in a novel intermediate feature space. Advantages resulting from the choice of feature space in our approach include robustness to differences in background texture (i.e., the method is not scene-dependent), larger discriminability between positive and negative samples, and simplification of the training process.","PeriodicalId":185674,"journal":{"name":"2016 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133131208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Edgar A. Bernal, Qun Li, Orhan Bulan, Wencheng Wu, S. Schweid
{"title":"Model-less and model-based computationally efficient motion estimation for video compression in transportation applications","authors":"Edgar A. Bernal, Qun Li, Orhan Bulan, Wencheng Wu, S. Schweid","doi":"10.1109/WACVW.2016.7470120","DOIUrl":"https://doi.org/10.1109/WACVW.2016.7470120","url":null,"abstract":"Block-based motion estimation is an important component in many video coding standards that aims at removing temporal redundancy between neighboring frames. Traditional methods for block-based motion estimation such as the Exhaustive Block Matching Algorithm (EBMA) are capable of achieving good matching performance but are computationally expensive. Alternatives to EBMA have been proposed to reduce the amount of search points by trading off matching optimality with computational resources. Although they exploit shared local spatial attributes around the target block, they fail to take advantage of the characteristics of the video sequences acquired with stationary cameras used in transportation and surveillance applications, where motion patterns are largely regularized; often, they also fail to yield semantically meaningful motion vector fields. In this paper, we propose two alternative approaches to improve the efficiency of motion estimation in video compression: (i) a highly efficient model-less approach that estimates the direction and magnitude of motion of objects in the scene and predicts the optimal search direction/neighborhood location for motion vectors; and (ii) a model-based approach that learns the dominant spatiotemporal characteristics of the motion patterns captured in the video via statistical models and enables reduced searches according to the constructed models. We demonstrate via experimental validation that the proposed methods attain computational savings, achieve improved reconstruction error and prediction capabilities for a given search neighborhood size, and yield more semantically meaningful motion vector fields when coupled with traditional motion estimation algorithms.","PeriodicalId":185674,"journal":{"name":"2016 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"74 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127237947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sejong Yoon, Mubbasir Kapadia, Pritish Sahu, V. Pavlovic
{"title":"Filling in the blanks: reconstructing microscopic crowd motion from multiple disparate noisy sensors","authors":"Sejong Yoon, Mubbasir Kapadia, Pritish Sahu, V. Pavlovic","doi":"10.1109/WACVW.2016.7470118","DOIUrl":"https://doi.org/10.1109/WACVW.2016.7470118","url":null,"abstract":"Tracking the movement of individuals in a crowd is an indispensable component to reconstructing crowd movement, with applications in crowd surveillance and data-driven animation. Typically, multiple sensors are distributed over wide area and often they have incomplete coverage of the area or the input introduces noise due to the tracking algorithm or hardware failure. In this paper, we propose a novel refinement method that complements existing crowd tracking solutions to reconstruct a holistic view of the microscopic movement of individuals in a crowd, from noisy tracked data with missing and even incomplete information. Central to our approach is a global optimization based trajectory estimation with modular objective functions. We empirically demonstrate the potential utility of our approach in various scenarios that are standard in crowd dynamic analysis and simulations.","PeriodicalId":185674,"journal":{"name":"2016 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127338440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Superpixels shape analysis for carried object detection","authors":"Blanca Delgado, Khalid Tahboub, E. Delp","doi":"10.1109/WACVW.2016.7470116","DOIUrl":"https://doi.org/10.1109/WACVW.2016.7470116","url":null,"abstract":"Video surveillance systems generate enormous amounts of data which makes the continuous monitoring of video a very difficult task. Re-identification of subjects in video surveillance systems plays a significant role in public safety. Recent work has focused on appearance modeling and distance learning to establish correspondence between images. However, real-life scenarios suggest that the majority of clothing worn tends to be non-discriminative. Attributes- based re-identification methods try to solve this problem by incorporating semantic attributes which are mid-level features learned a prior. In this paper we present a framework to recognize attributes with applications to carried objects detection. We present a supervised approach based on the contours and shapes of superpixels and histogram of oriented gradients. An experimental evaluation is described using a dataset that was recorded at the Greater Cleveland Regional Transit Authority.","PeriodicalId":185674,"journal":{"name":"2016 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130814781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Swetha, Anand Mishra, Guruprasad M. Hegde, C. V. Jawahar
{"title":"Efficient object annotation for surveillance and automotive applications","authors":"S. Swetha, Anand Mishra, Guruprasad M. Hegde, C. V. Jawahar","doi":"10.1109/WACVW.2016.7470117","DOIUrl":"https://doi.org/10.1109/WACVW.2016.7470117","url":null,"abstract":"Accurately annotated large video data is critical for the development of reliable surveillance and automotive related vision solutions. In this work, we propose an efficient and yet accurate annotation scheme for objects in videos (pedestrians in this case) with minimal supervision. We annotate objects with tight bounding boxes. We propagate the annotations across the frames with a self training based approach. An energy minimization scheme for the segmentation is the central component of our method. Unlike the popular grab cut like segmentation schemes, we demand minimal user intervention. Since our annotation is built on an accurate segmentation, our bounding boxes are tight. We validate the performance of our approach on multiple publicly available datasets.","PeriodicalId":185674,"journal":{"name":"2016 IEEE Winter Applications of Computer Vision Workshops (WACVW)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116937545","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}