{"title":"Hallucination space relationship learning to improve very low resolution face recognition","authors":"Juhyun Ahn, Daijin Kim, S. I. Ch'ng","doi":"10.1109/ACPR.2015.7486455","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486455","url":null,"abstract":"It is known that face recognition rate is affected by the resolution of probe image. Therefore it is natural to expect improving the resolution via face hallucination would increase the recognition rate. However, it was concluded in the previous works that improvement in the visual quality does not necessarily lead to a better recognition rate, and performance of face hallucination for recognition is especially poor when very low resolution (VLR) images are used. Experiment results in this paper will show that there is a correlation between visual quality and recognition rate. And hallucination space relationship learning algorithm is proposed which is robust on hallucinating VLR images to improve the performance in terms of both visual quality and recognition rate.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121856867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fine pose estimation of known objects in cluttered scene images","authors":"Sudipto Banerjee, Sanchit Aggarwal, A. Namboodiri","doi":"10.1109/ACPR.2015.7486579","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486579","url":null,"abstract":"Understanding the precise 3D structure of an environment is one of the fundamental goals of computer vision and is challenging due to a variety of factors such as appearance variation, illumination, pose, noise, occlusion and scene clutter. A generic solution to the problem is ill-posed due to the loss of depth information during imaging. In this paper, we consider a specific but common situation, where the scene contains known objects. Given 3D models of a set of known objects and a cluttered scene image, we try to detect these objects in the image, and align 3D models to their images to find their exact pose. We develop an approach that poses this as a 3D-to-2D alignment problem. We also deal with pose estimation of 3D articulated objects in images. We evaluate our proposed method on BigBird dataset and our own tabletop dataset, and present experimental comparisons with state-of-the-art methods.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117141411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Efficient objectness via saliency seeds and contour segments","authors":"Rigen Te, Cheng Yan","doi":"10.1109/ACPR.2015.7486613","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486613","url":null,"abstract":"Object proposal is a new paradigm for improving efficiency for object detection. We propose an efficient method for object proposals by saliency seeds and contour segments. A simple saliency method is used to get several salient seeds in the image to target all the probable objects appeared in image, roughly leaving background regions out of consideration. Then we further score each of the salient seeds by using a bounding box strategy. If the bounding box contains more contour segments of the seed, it is assumed to be the object proposal more strongly. For efficiency, we utilize Pair of Adjacent Segments (PAS) as the contour segment feature, which is easy to detect and can describe the location and scale of contours compactly. After getting the proposal regions, those PAS features are also used for classification task. Experiments show that the proposed method is very effective. It has achieved comparable result to state of the art methods with higher efficiency and also provide auxiliary information to later classification step.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128682558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multi-attribute learning for pedestrian attribute recognition in surveillance scenarios","authors":"Dangwei Li, Xiaotang Chen, Kaiqi Huang","doi":"10.1109/ACPR.2015.7486476","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486476","url":null,"abstract":"In real video surveillance scenarios, visual pedestrian attributes, such as gender, backpack, clothes types, are very important for pedestrian retrieval and person reidentification. Existing methods for attributes recognition have two drawbacks: (a) handcrafted features (e.g. color histograms, local binary patterns) cannot cope well with the difficulty of real video surveillance scenarios; (b) the relationship among pedestrian attributes is ignored. To address the two drawbacks, we propose two deep learning based models to recognize pedestrian attributes. On the one hand, each attribute is treated as an independent component and the deep learning based single attribute recognition model (DeepSAR) is proposed to recognize each attribute one by one. On the other hand, to exploit the relationship among attributes, the deep learning framework which recognizes multiple attributes jointly (DeepMAR) is proposed. In the DeepMAR, one attribute can contribute to the representation of other attributes. For example, the gender of woman can contribute to the representation oflong hair and wearing skirt. Experiments on recent popular pedestrian attribute datasets illustrate that our proposed models achieve the state-of-the-art results.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125307728","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Chunfeng Song, Yongzhen Huang, Zhenyu Wang, Liang Wang
{"title":"1000fps human segmentation with deep convolutional neural networks","authors":"Chunfeng Song, Yongzhen Huang, Zhenyu Wang, Liang Wang","doi":"10.1109/ACPR.2015.7486548","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486548","url":null,"abstract":"Efficiency and effectiveness are two key factors to evaluate a human segmentation algorithm for real vision applications. However, most existing algorithms only focus on one of them. That is, fast and accurate human segmentation is not yet well addressed. In this paper, we propose a super-fast and highly accurate human segmentation method with very deep convolutional neural networks. We also provide a comprehensive study on the proposed approach, including different net structures, various techniques of alleviating over-fitting, and performance enhancement with different extra data. Experimental results on the database of Baidu people segmentation competition [1] demonstrate that the proposed model outperforms traditional segmentation algorithms in accuracy and speed. Although it is slightly worse than the very complex champion algorithm, it is encouraging that our method can obtain more than 10,000 times acceleration, showing that it has great potential for practical applications.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"131 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128712825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Tsuchiya, Yuji Yamauchi, Takayoshi Yamashita, H. Fujiyoshi
{"title":"Transfer forest based on covariate shift","authors":"M. Tsuchiya, Yuji Yamauchi, Takayoshi Yamashita, H. Fujiyoshi","doi":"10.1109/ACPR.2015.7486605","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486605","url":null,"abstract":"Random Forest, a multi-class classifier based on statistical learning, is widely used in applications because of its high generalization performance due to randomness. However, in applications such as object detection, disparities in the distributions of the training and test samples from the target scene are often inevitable, resulting in degraded performance. In this case, the training samples need to be reacquired for the target scene, typically at a very high human acquisition cost. To solve this problem, transfer learning has been proposed. In this paper, we present data-level transfer learning for a Random Forest using covariate shift. Experimental results show that the proposed method, called Transfer Forest, can adapt to the target domain by transferring training samples from an auxiliary domain.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128774751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A novel fuzzy LBP based symbolic representation technique for classification of medicinal plants","authors":"Y. Naresh, H. S. Nagendraswamy","doi":"10.1109/ACPR.2015.7486558","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486558","url":null,"abstract":"In this paper, a novel fuzzy LBP model for extracting texture features from medicinal plant leaves is proposed. The proposed method is invariant to image transformations and independent of any threshold. Concept of hierarchical clustering based on inconsistency coefficient is used to produce natural clusters for a particular species capturing intra-class variations due to environmental conditions and acquisition system. Interval valued type symbolic feature vector is used to represent each cluster effectively. Thus the proposed system suggests choosing multiple representatives for each species to make the representation more effective and robust. A chi-square distance measure is used to establish matching between the test and reference feature vectors of plant leaves and a nearest neighbor classification technique is used to classify an unknown test sample of medicinal plant leaf. Extensive experiments are conducted to demonstrate the efficacy of the proposed model on our own data set and other publically available leaf datasets. Results of the proposed work has been compared with the contemporary work and found to be superior.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128850027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fine-grain annotation of cricket videos","authors":"R. Sharma, K. Sankar, C. V. Jawahar","doi":"10.1109/ACPR.2015.7486538","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486538","url":null,"abstract":"The recognition of human activities is one of the key problems in video understanding. Action recognition is challenging even for specific categories of videos, such as sports, that contain only a small set of actions. Interestingly, sports videos are accompanied by detailed commentaries available online, which could be used to perform action annotation in a weakly-supervised setting. For the specific case of Cricket videos, we address the challenge of temporal segmentation and annotation of actions with semantic descriptions. Our solution consists of two stages. In the first stage, the video is segmented into \"scenes\", by utilizing the scene category information extracted from text-commentary. The second stage consists of classifying videoshots as well as the phrases in the textual description into various categories. The relevant phrases are then suitably mapped to the video-shots. The novel aspect of this work is the fine temporal scale at which semantic information is assigned to the video. As a result of our approach, we enable retrieval of specific actions that last only a few seconds, from several hours of video. This solution yields a large number of labelled exemplars, with no manual effort, that could be used by machine learning algorithms to learn complex actions.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121708556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Lingxiao Song, Man Zhang, Qi Li, Zhenan Sun, R. He
{"title":"Float greedy-search-based subspace clustering","authors":"Lingxiao Song, Man Zhang, Qi Li, Zhenan Sun, R. He","doi":"10.1109/ACPR.2015.7486494","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486494","url":null,"abstract":"Many kinds of efficient greedy subspace clustering methods have been proposed to cut down the computation time in clustering large-scale multimedia datasets. However, these methods are easy to fall into local optimum due to the inherent characteristic of greedy algorithms, which are step-optimal only. To alleviate this problem, this paper proposes a novel greedy subspace clustering strategy based on floating search, called Float Greedy Subspace Clustering (FloatGSC). In order to control the complexity, the nearest subspace neighbor is added in a greedy way, and the subspace is updated by adding an orthogonal basis involved with the newly added data points in each iteration. Besides, a backtracking mechanism is introduced after each iteration to reject wrong neighbors selected in previous iterations. Extensive experiments on motion segmentation and face clustering show that our algorithm can significantly improve the clustering accuracy without sacrificing much computational time, compared with previous greedy subspace clustering methods.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125734825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yi Li, Yanqing Guo, J. Guo, Ming Li, Xiangwei Kong
{"title":"CRF with locality-consistent dictionary learning for semantic segmentation","authors":"Yi Li, Yanqing Guo, J. Guo, Ming Li, Xiangwei Kong","doi":"10.1109/ACPR.2015.7486555","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486555","url":null,"abstract":"The use of top-down categorization information in bottom-up semantic segmentation can significantly improve its performance. The basic Conditional Random Field (CR-F) model can capture the local contexture information, while the locality-consistent sparse representation can obtain the category-level priors and the relationship infeature space. In this paper, we propose a novel semantic segmentation method based on an innovative CRF with locality-consistent dictionary learning. The framework aims to model the local structure in both location and feature space as well as encourage the discrimination of dictionary. Moreover, an adapted algorithm for the proposed model is described. Extensive experimental results on Graz-02, PASCAL VOC 2010 and MSRC-21 databases demonstrate that our method is comparable to or outperforms state-of-the-art Bag-of-Features (BoF) based segmentation methods.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125958625","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}