{"title":"Novel Keypoint Registration for Fast and Robust Pose Detection on Mobile Phones","authors":"Tatsuya Kobayashi, H. Kato, H. Yanagihara","doi":"10.1109/ACPR.2013.67","DOIUrl":"https://doi.org/10.1109/ACPR.2013.67","url":null,"abstract":"We present a novel vision-based pose detection method that can be used in mobile AR services. Conventional methods are unable to meet all the requirements such as complexity, robustness and memory consumption for mobile AR services because of their trade-off relationship. In this paper, we propose a novel key point registration approach to solve the problem. Our registration method detects key point candidates and their binary descriptors from a small number of essential training images to improve robustness to changes in viewpoint. The detected features are screened by our two-stage selection method that selects only good features for pose detection. Experimental results demonstrate that our approach both improves the robustness of the conventional method by about 50% and speeds up runtime processing by about 7-10% with small memory consumption.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124769548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Brahmastro Kresnaraman, Y. Mekada, Tomokazu Takahashi, H. Murase
{"title":"Learning Based Reconstruction of Grayscale Face Image from Far-Infrared Image","authors":"Brahmastro Kresnaraman, Y. Mekada, Tomokazu Takahashi, H. Murase","doi":"10.1109/ACPR.2013.74","DOIUrl":"https://doi.org/10.1109/ACPR.2013.74","url":null,"abstract":"It is important for security surveillance systems to operate continuously for 24 hours. During the night, use of far-infrared cameras is preferable in outdoor situations due to a number of reasons. However, the person in the image is often unrecognizable. This paper attempts to estimate the face from his/her far-infrared image. The estimation is done through two phases, a holistic estimation and a patch based one. In each of these phases, a learning based approach is employed, which learns the relationship between grayscale and far-infrared face images from pairs of the images of a large number of persons. Canonical Correlation Analysis (CCA) is performed to obtain the maximum correlation in the data. Locally Linear Embedding (LLE) is performed to estimate grayscale face image. Three types of experiments were done with this method and evaluated by PSNR. These experiments show a good result in estimating face image whose face images of different expressions were included in training data.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129300574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Masatoshi Ando, Kanji Tanaka, Yousuke Inagaki, Yuuto Chokushi, Shogo Hanada
{"title":"Common Landmark Discovery for Object-Level View Image Retrieval: Modeling and Matching of Scenes via Bag-of-Bounding-Boxes","authors":"Masatoshi Ando, Kanji Tanaka, Yousuke Inagaki, Yuuto Chokushi, Shogo Hanada","doi":"10.1109/ACPR.2013.19","DOIUrl":"https://doi.org/10.1109/ACPR.2013.19","url":null,"abstract":"Object-level view image retrieval for robot vision applications has been actively studied recently, as they can provide semantic and compact method for efficient scene matching. In existing frameworks, landmark objects are extracted from an input view image by a pool of pretrained object detectors, and used as an image representation. To improve the compactness and autonomy of object-level view image retrieval, we here present a novel method called ``common landmark discovery\". Under this method, landmark objects are mined through common pattern discovery (CPD) between an input image and known reference images. This approach has three distinct advantages. First, the CPD-based object detection is unsupervised, and does not require pretrained object detector. Second, the method attempts to find fewer and larger object patterns, which leads to a compact and semantically robust view image descriptor. Third, the scene matching problem is efficiently solved as a lower-dimensional problem of computing region overlaps between landmark objects, using a compact image representation in a form of bag-of-bounding-boxes (BoBB).","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117144640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Performance Evaluation of Image Feature Detectors and Descriptors for Outdoor-Scene Visual Navigation","authors":"Dzulfahmi, N. Ohta","doi":"10.1109/ACPR.2013.159","DOIUrl":"https://doi.org/10.1109/ACPR.2013.159","url":null,"abstract":"Scene image matching is often used for positioning of a visually navigated autonomous robot. The robot memorizes the scene as an image at each navigation point in the teaching mode, and knows being at the same position when the outside scene is matched to the image in the playback mode. The scene matching is usually accomplished by feature-based image matching methods, such as SIFT or SURF. However the problem is that matching results of such methods are greatly affected by changes in illumination condition. Therefore, it is important to know which method is robust to the illumination change. Several performance evaluation results of these matching methods have been reported, but they are not focusing on illumination change problem. In this paper, we present performance comparison results of these feature-based image matching methods against illumination change in outdoor scenes assuming usage for visual navigation purpose. We also encounter another problem when conducting such the comparison for visual navigation. In this application, the matching score gradually increases as approaching the matching point, and gradually decreases as being apart from that point. This impedes to define the right matching response (ground truth). We present one method by which giving the right response.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115606512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ramachandra Raghavendra, K. Raja, Bian Yang, C. Busch
{"title":"Combining Iris and Periocular Recognition Using Light Field Camera","authors":"Ramachandra Raghavendra, K. Raja, Bian Yang, C. Busch","doi":"10.1109/ACPR.2013.22","DOIUrl":"https://doi.org/10.1109/ACPR.2013.22","url":null,"abstract":"Iris and Periocular biometrics has proved its effectiveness in accurately verifying the subject of interest. Recent improvements in visible spectrum Iris and Periocular verification have further boosted its application to unconstrained scenarios. However existing visible Iris verification systems suffer from low quality samples because of the limited depth-of-field exhibited by the conventional Iris capture systems. In this work, we propose a robust Iris and Periocular erification scheme in visible spectrum using Light Field Camera (LFC). Since the light field camera can provide multiple focus images in single capture, we are motivated to investigate its applicability for robust Iris and Periocular verification by exploring its all-in-focus property. Further, the use of all-in-focus property will extend the depth-of-focus and overcome the problem of focus that plays a predominant role in robust Iris and Periocular verification. We first collect a new Iris and Periocular biometric database using both light field and conventional camera by simulating real life scenarios. We then propose a new scheme for feature extraction and classification by exploring the combination of Local Binary Patterns (LBP) and Sparse Reconstruction Classifier (SRC). Extensive experiments are carried out on the newly collected database to bring out the merits and demerits on applicability of light field camera for Iris and Periocular verification. Finally, we also present the results on combining the information from Iris and Periocular biometrics using weighted sum rule.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"193 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122546824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust Angle Invariant 1D Barcode Detection","authors":"Alessandro Zamberletti, I. Gallo, S. Albertini","doi":"10.1109/ACPR.2013.17","DOIUrl":"https://doi.org/10.1109/ACPR.2013.17","url":null,"abstract":"Barcode reading mobile applications that identify products from pictures taken using mobile devices are widely used by customers to perform online price comparisons or to access reviews written by others. Most of the currently available barcode reading approaches focus on decoding degraded barcodes and treat the underlying barcode detection task as a side problem that can be addressed using appropriate object detection methods. However, the majority of modern mobile devices do not meet the minimum working requirements of complex general purpose object detection algorithms and most of the efficient specifically designed barcode detection algorithms require user interaction to work properly. In this paper, we present a novel method for barcode detection in camera captured images based on a supervised machine learning algorithm that identifies one-dimensional barcodes in the two-dimensional Hough Transform space. Our model is angle invariant, requires no user interaction and can be executed on a modern mobile device. It achieves excellent results for two standard one-dimensional barcode datasets: WWU Muenster Barcode Database and ArTe-Lab 1D Medium Barcode Dataset. Moreover, we prove that it is possible to enhance the overall performance of a state-of-the-art barcode reading algorithm by combining it with our detection method.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123224973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Traffic Sign Recognition Using Complementary Features","authors":"Suisui Tang, Lin-Lin Huang","doi":"10.1109/ACPR.2013.63","DOIUrl":"https://doi.org/10.1109/ACPR.2013.63","url":null,"abstract":"Traffic sign recognition is difficult due to the low resolution of image, illumination variation and shape distortion. On the public dataset GTSRB, the state-of-the-art performance have been obtained by convolutional neural networks (CNNs), which learn discriminative features automatically to achieve high accuracy but suffer from high computation costs in both training and classification. In this paper, we propose an effective traffic sign recognition method using multiple features which have demonstrated effective in computer vision and are computationally efficient. The extracted features are the histogram of oriented gradients (HOG) feature, Gabor filter feature and local binary pattern (LBP) feature. Using a linear support vector machine (SVM) for classification, each feature yields fairly high accuracy. The combination of three features has shown good complementariness and yielded competitively high accuracy. On the GTSRB dataset, our method reports an accuracy of 98.65%.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124228517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Zhizhen Liang, Shixiong Xia, Jin Liu, Yong Zhou, Lei Zhang
{"title":"A Majorization-Minimization Approach to Lq Norm Multiple Kernel Learning","authors":"Zhizhen Liang, Shixiong Xia, Jin Liu, Yong Zhou, Lei Zhang","doi":"10.1109/ACPR.2013.54","DOIUrl":"https://doi.org/10.1109/ACPR.2013.54","url":null,"abstract":"Multiple kernel learning (MKL) usually searches for linear (nonlinear) combinations of predefined kernels by optimizing some performance measures. However, previous MKL algorithms cannot deal with Lq norm MKL if q<;1 due to the non-convexity of Lq (q<;1) norm. In order to address this problem, we apply a majorization-minimization approach to solve Lq norm MKL in this paper. It is noted that the proposed method only involves solving a series of support vector machine problems, which makes the proposed method simple and effective. We also theoretically demonstrate that the limit points of the sequence generated from our iterative scheme are stationary points of the optimization problem under proper conditions. Experiments on synthetic data and some benchmark data sets, and gene data sets are carried out to show the effectiveness of the proposed method.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131672559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
D. Muramatsu, Yasushi Makihara, Haruyuki Iwama, Takuya Tanoue, Y. Yagi
{"title":"Gait Verification System for Supporting Criminal Investigation","authors":"D. Muramatsu, Yasushi Makihara, Haruyuki Iwama, Takuya Tanoue, Y. Yagi","doi":"10.1109/ACPR.2013.195","DOIUrl":"https://doi.org/10.1109/ACPR.2013.195","url":null,"abstract":"We constructed gait verification system for criminal investigation. The system is designed so that criminal investigators can use it and obtain professional gait verification results. We think the system can support criminal investigation where gait can be a clue of the perpetrator's identity. We summarize the constructed system in this paper.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131997013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Set-Based Feature Learning for Person Re-identification via Third-Party Images","authors":"Yanna Zhao, Lei Wang, Yuncai Liu","doi":"10.1109/ACPR.2013.87","DOIUrl":"https://doi.org/10.1109/ACPR.2013.87","url":null,"abstract":"Person re-identification from disjoint camera views has been an important and unsolved problem due to large variations in illumination, viewpoint and pose. One way to attack this is by designing a new, more powerful image representation. However, we believe that existing representations are already sufficient. The main difficulty is how to pick the most informative information using these representations. Inspired by the prototype theory from the cognition field and Exemplar-SVM, we propose a novel and simple set-based feature learning re-identification method via third-party images. In our settings, each query/gallery example is an image set of the same individual, not just a single image. Discriminative features of a certain individual image set are explored from the third-party images. Comparisons with state-of-the-art methods on benchmark datasets demonstrate impressive results using simple and common features.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125563702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}