Rui Wang, Joseph A. Camilo, L. Collins, Kyle Bradbury, Jordan M. Malof
{"title":"The poor generalization of deep convolutional networks to aerial imagery from new geographic locations: an empirical study with solar array detection","authors":"Rui Wang, Joseph A. Camilo, L. Collins, Kyle Bradbury, Jordan M. Malof","doi":"10.1109/AIPR.2017.8457965","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457965","url":null,"abstract":"Convolutional neural networks (CNNs) have recently achieved unprecedented performance for the automatic recognition of objects (e.g., buildings, roads, or vehicles) in color aerial imagery. Although these results are promising, questions remain about their practical applicability. This is because there is a wide variability in the visual characteristics of remote sensing imagery across different geographic locations, and CNNs are often trained and tested on imagery from nearby (or the same) geographic locations. It is therefore unclear whether trained CNNs will perform well on new, previously unseen, geographic locations, which is an important practical consideration. In this work we investigate this problem when applying CNNs for solar array detection on a large aerial imagery dataset comprised of two nearby US cities. We compare the performance of CNNs under two conditions: training and testing on the same city vs training on one city and testing on another city. We discuss several subtle difficulties with these experiments and make recommendations. We show that there can be substantial performance loss in second case, when compared to the first. We also investigate how much training data is required from the unseen city in order to fine-tune the CNN so that it performs well. We investigate several different fine-tuning strategies, yielding a clear winner.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115570121","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Benchmarking Convolutional Neural Networks for Object Segmentation and Pose Estimation","authors":"T. Le, L. Hamilton, A. Torralba","doi":"10.1109/AIPR.2017.8457943","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457943","url":null,"abstract":"Convolutional neural networks (CNNs), particularly those designed for object segmentation and pose estimation, are now applied to robotics applications involving mobile manipulation. For these robotic applications to be successful, robust and accurate performance from the CNNs is critical. Therefore, in order to develop an understanding of CNN performance, several CNN architectures are benchmarked on a set of metrics for object segmentation and pose estimation. This paper presents these benchmarking results, which show that metric performance is dependent on the complexity of network architectures. These findings can be used to guide and improve the development of CNNs for object segmentation and pose estimation in the future.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116842960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}