A. Shafiekhani, Arun Prabhu Dhanapal, J. Gillman, F. Fritschi, G. DeSouza
{"title":"Automated Classification of Wrinkle Levels in Seed Coat Using Relevance Vector Machine","authors":"A. Shafiekhani, Arun Prabhu Dhanapal, J. Gillman, F. Fritschi, G. DeSouza","doi":"10.1109/AIPR.2017.8457939","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457939","url":null,"abstract":"Seed-coat wrinkling in soybean is often observed when seeds are produced in adverse environmental conditions and it has been associated with low germinability. Manually rating seeds is time consuming, error prone and fatiguing - leading to even more errors. In this paper, an automated approach for the rating of seed-coat wrinkling using computer vision and machine learning algorithms is presented. The proposed system provides a GUI for ground truth annotation and a pipeline consisting of seed segmentation, feature extraction and classification using multi-class Relevance Vector Machines (mRVM). This research also proposes a reliable new feature for seed-coat rating based on texture. An additional contribution of this paper is a database of annotated seed images, which is being made available to researchers in the field. The results showed an accuracy in wrinkling rating of 86 % for matches within ± 1 scores from the ground truth.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127207343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. F. Said, Vinay Kashyap, Namrata Choudhury, F. Akhbari
{"title":"A Cost-Effective, Fast, and Robust Annotation Tool","authors":"A. F. Said, Vinay Kashyap, Namrata Choudhury, F. Akhbari","doi":"10.1109/AIPR.2017.8457958","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457958","url":null,"abstract":"Deep learning requires huge datasets for training their models. In most cases, dataset generation is done manually or by using conventional approaches which are time consuming, costly, and prone to inaccuracy. There is an urgent need for developing fast, robust, and effective techniques to annotate each image with different labels and classes based on its contents. In this paper, we propose the use of in-house annotation tool that can be used for generating accurate datasets in a short time with minimal human interference. The proposed tool reads a captured video and the operator manually highlights the objects with their assigned classes in the initial frame. A robust and fast object tracking approach was developed to detect and track each highlighted object for subsequent frames. Experimental results using our annotation tool show 50x to 200x faster processing time compared to conventional annotation methods in addition to better performance with higher accuracy.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"280 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116545170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Scene wireframes sketching for UAVs","authors":"R. Santos, X. López, Xosé R. Fernández-Vidal","doi":"10.1109/AIPR.2017.8457938","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457938","url":null,"abstract":"This paper introduces novel insights to improve the state of the art line-based unsupervised observation and abstraction models of urban environments. The scene observation is performed by an UAV, using self-detected and matched straight segments from streamed video frames. The increasing use of autonomous UAV s inside buildings and human built structures demands new accurate and comprehensive representations for their environment. Most of the 3D scene abstraction methods published are using invariant feature point matching, nevertheless some sparse 3D point clouds do not concisely represent the structure of the environment. Likewise, line clouds constructed by short and redundant segments with unaccurate directions will limit the understanding of the objective scenes, that include environments with no texture, or whose texture resembles a repetitive pattern. The presented approach is based on observation and representation models using the straight line segments, whose resemble the limits of an urban indoor or outdoor environment. The goal of the work is to get a better 3D representation for future autonomous UAV.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"2 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114042578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Video-Level Binocular Tone-mapping Framework Based on Temporal Coherency Algorithm","authors":"Mingyue Feng, M. Loew","doi":"10.1109/AIPR.2017.8457950","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457950","url":null,"abstract":"A binocular tone-mapping framework can generate a binocular low-dynamic range (LDR) image pair that preserves more human-perceivable visual contents than a single LDR image. In this paper, to solve the temporal coherency problem when extending from this image-level system to a video-level system, we proposed a new binocular framework that integrates the existing image-level framework and the temporal coherency algorithm. The experimental data show that this proposed new framework can effectively solve the temporal coherency problem and generate binocular LDR videos without disturbing effects.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121925554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Remote Sensing to Analyze Wealth, Poverty, and Crime","authors":"J. Irvine, Richard J. Wood, Payden Mcbee","doi":"10.1109/AIPR.2017.8457971","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457971","url":null,"abstract":"understanding of economic, social, and cultural characteristics of a society is critical to effective government policy and successful commercial undertakings. Obtaining this information, however, often requires direct interactions with the local populace through surveys or other costly methods. We address this challenge by combining automated processing of satellite imagery with advanced modeling techniques. We have developed methods for inferring measures of wellbeing and perceptions of crime from commercial satellite imagery. Through analysis of commercial satellite imagery and coincident survey data, previous research has demonstrated models for rural afghanistan and selected countries in sub-saharan africa. The findings show the potential for predicting peoples' attitudes about the a variety of social, economic, and political issues, based only on the imagery-derived information. This paper extends the previous research, focusing on wealth, poverty, and crime. We present models to predict indicators and quantify model performance through cross-validation. The paper concludes with recommendations for future exploration.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"36 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132552875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust and Accurate Objects Measurement in Real-World Based on Camera System","authors":"A. F. Said","doi":"10.1109/AIPR.2017.8457954","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457954","url":null,"abstract":"Object's dimension and its proximity in real-world plays a critical role in safe navigation and collision avoidance in autonomous cars. An accurate, reliable, and cost-effective approach was developed in this paper to measure the object's dimension (distance, width, and height) in real-world solely based on camera system. Mathematical representations were derived to accurately measure object dimensions and extract extrinsic camera parameters while driving. Giving the bounding box coordinates around each object in the captured frame, the proposed approach automatically and accurately measure the object's dimension in real-world (ft.) instead of pixels. The derived models were verified and tested against the ground truth data which showed strong correlation.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125197753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Zakariya A. Oraibi, Morgane Irio, A. Hafiane, K. Palaniappan
{"title":"Texture Classification using Multiple Local Descriptors","authors":"Zakariya A. Oraibi, Morgane Irio, A. Hafiane, K. Palaniappan","doi":"10.1109/AIPR.2017.8457968","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457968","url":null,"abstract":"Classifying images based on texture features is an active topic in computer vision and pattern recognition field. Many applications like biomedical image analysis, image retrieval, and face recognition emerged from texture classification task. In this paper, we propose a new method to classify texture images by combining multiple histogram-based texture descriptors. First, we compute new efficient features called Joint Motif Labels (JML) and Motif Patterns (MP) descriptors. Both descriptors are based on the motif Peano scan concept that traverses image pixels in a 2×2 neighborhood producing one of 12 motif patterns, according to certain criteria. JML uses additional information, mean and variance, as joint distribution with motif patterns. After that, texture descriptors like Rotation Invariance Co-occurrence Among Local Binary Pattern (RIC-LBP) and Joint Adaptive Median Binary Pattern (JAMBP) are combined along with the new JML and MP descriptors in order to improve the classification performance. Experiments are performed on challenging texture datasets namely, KTH- TIPS-2b and DTD using two classifiers, kNN and SVM. The experiments demonstrate that our approach performs better than the single best texture descriptor with an accuracy of 67.2% and 43.5% on both datasets respectively.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"160 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133582623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Gaël Mondonneix, S. Chabrier, Jean-Martial Mari, A. Gabillon
{"title":"Tahitian Pearls' Luster Assessment Automation","authors":"Gaël Mondonneix, S. Chabrier, Jean-Martial Mari, A. Gabillon","doi":"10.1109/AIPR.2017.8457974","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457974","url":null,"abstract":"Luster assessment stands at the crossroads of different fields and there is very few literature specifically dedicated to it. In a perspective of automating culture pearls' luster assessment, a way to extract features out of pearls' photographs is proposed and tested on a real dataset labeled by a human expert. After training, an SVM using these features can predict luster quality of new pearls with up to 87.3 % (± 5.7) accuracy. Moreover, it turns out that some of these features could be used for developing an objective luster quality control.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130260152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Restoration of Medical Images Using Genetic Algorithms","authors":"A. Sheta","doi":"10.1109/AIPR.2017.8457940","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457940","url":null,"abstract":"Image restoration is still one of the most important areas of medical image processing. Image restoration concerns about the removal or reduction of degradations in an image that could happen during the acquisition process. Being able to restore a medical image helps providing a better diagnosis and treatment. One of the most common blurring is the motion blur. Many restoration algorithms were proposed to solve the image restoration problem such as Wiener Filter, Lucy-Richardson and Blind Deconvolution Algorithms. These algorithms have varied performance, computational complexity, and abilities to deal with noisy images. They also require the knowledge of the Point Spread function (PSF) such that image deconvolution can be implemented. Restoration of an image is extremely reliant on the quality of the estimation technique used to find an accurate PSF parameters (i.e. motion length and motion angle). In this paper, we adopt Genetic Algorithms (GAs) to find the optimal PSF parameters such that a Wiener filter can be used for image restoration. We adopted number of statistical evaluation criteria to asses the quality of our proposed method. We applied our method on a number of medical images with various additive Gaussian noise. The developed results show that our proposed algorithm, PSF generated by GAs, is showing better results compared to other known methods in the literature in the absence of the real PSF.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"92 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121326847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Benchmarking Convolutional Neural Networks for Object Segmentation and Pose Estimation","authors":"T. Le, L. Hamilton, A. Torralba","doi":"10.1109/AIPR.2017.8457942","DOIUrl":"https://doi.org/10.1109/AIPR.2017.8457942","url":null,"abstract":"Convolutional neural networks (CNNs), particularly those designed for object segmentation and pose estimation, are now applied to robotics applications involving mobile manipulation. For these robotic applications to be successful, robust and accurate performance from the CNNs is critical. Therefore, in order to develop an understanding of CNN performance, several CNN architectures are benchmarked on a set of metrics for object segmentation and pose estimation. This paper presents these benchmarking results, which show that metric performance is dependent on the complexity of network architectures. These findings can be used to guide and improve the development of CNNs for object segmentation and pose estimation in the future.","PeriodicalId":128779,"journal":{"name":"2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"26 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114013603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}