2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)最新文献

筛选
英文 中文
Iris biometrie: Is the near-infrared spectrum always the best? 虹膜生物识别:近红外光谱总是最好的吗?
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486616
Mohammed A. M. Abdullah, J. Chambers, W. L. Woo, S. Dlay
{"title":"Iris biometrie: Is the near-infrared spectrum always the best?","authors":"Mohammed A. M. Abdullah, J. Chambers, W. L. Woo, S. Dlay","doi":"10.1109/ACPR.2015.7486616","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486616","url":null,"abstract":"Previous work on iris recognition focused on either Visible Light (VL), Near-Infrared (NIR) imaging or the fusion between them. However, limited numbers of works have compared the iris biometric performance under both VL and NIR spectrum using images taken from the same subject. In this paper, we explore the differences in iris recognition performance across the VL and NIR spectrum. In addition, we investigate the possibility of cross-channel matching between the VL and NIR imaging. We carried out our experiments on the UTIRIS database which contains iris images taken under both the VL and NIR spectrum for the same subject. This paper is amongst the first studies which compares the performance of iris biometric under the NIR and VL spectrum and produces comparative experiments between these types of data. Experimental results indicate that the VL and NIR images provide complementary features for the iris pattern and their fusion improves the recognition performance. In addition, the experiments indicate that cross-channel matching between VL and NIR images is feasible.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128203319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Offline handwritten Devanagari word recognition: Information fusion at feature and classifier levels 离线手写德文字识别:特征和分类器层次的信息融合
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486597
Bikash Shaw, U. Bhattacharya, S. K. Parui
{"title":"Offline handwritten Devanagari word recognition: Information fusion at feature and classifier levels","authors":"Bikash Shaw, U. Bhattacharya, S. K. Parui","doi":"10.1109/ACPR.2015.7486597","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486597","url":null,"abstract":"This article presents our recent study on fusion of information at feature and classifier output levels for improved performance of offline handwritten Devanagari word recognition. We consider here two state-of-the-art features, viz., Directional Distance Distribution (DDD) and Gradient-Structural-Concavity (GSC) features along with multi-class SVM classifiers. Here, we study various combinations of DDD features along with one or more features from the GSC feature set. We experiment by presenting different combined feature vectors as input to SVM classifiers. Also, the output vectors of different SVM classifiers fed with different feature vectors are combined by another SVM classifier. The combination of the outputs of two SVMs each being fed with a different feature vector provides superior performance to the performance of a single SVM classifier fed with the combined feature vector. Experimental results are obtained on a large handwritten Devanagari word sample image database of 100 Indian town names. The recognition results on its test samples show that SVM recognition output of DDD features combined with the SVM output of GSC features improves the final recognition accuracy significantly.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115034751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Privacy-conscious human detection using low-resolution video 使用低分辨率视频的具有隐私意识的人类检测
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486519
Nobuhiro Miyazaki, K. Tsuji, Mingxie Zheng, M. Nakashima, Yuji Matsuda, E. Segawa
{"title":"Privacy-conscious human detection using low-resolution video","authors":"Nobuhiro Miyazaki, K. Tsuji, Mingxie Zheng, M. Nakashima, Yuji Matsuda, E. Segawa","doi":"10.1109/ACPR.2015.7486519","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486519","url":null,"abstract":"We propose a human detection method for obtaining human flow information from low-resolution video generated by existing surveillance cameras. To use cameras in public spaces, it is necessary to protect the privacy of individuals appearing in the videos. We use low-resolution video in which individuals cannot be identified. In low-resolution video, human detection is more difficult than in high-resolution video because the human region consists of few pixels and has little information. Furthermore, it is challenging to detect an individual when others are captured nearby or occluded the individual. The proposed method offers detection based on the shape of the head, which is kept in a low-resolution image. In addition, to reduce false positives, which occur when detecting heads with simple shapes, head candidates are verified using the shape of the upper-body. The experimental results indicate a detection rate higher than 70% when the head width is from three to eight pixels (corresponding to about 20 to 50 pixels of the human height) in 428 people.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134202963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
WA-ICP algorithm for tackling ambiguous correspondence 用于处理模糊通信的WA-ICP算法
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486469
Rong Wang, Z. Geng
{"title":"WA-ICP algorithm for tackling ambiguous correspondence","authors":"Rong Wang, Z. Geng","doi":"10.1109/ACPR.2015.7486469","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486469","url":null,"abstract":"This paper proposes a weighted average iterative closest point (ICP) algorithm (WA-ICP) to register multiple 3D surfaces. The new algorithm aims to tackle ambiguous correspondence and clearly improves the efficiency and reliability of 3D surface registration. We developed the WA-ICP theory, implemented the algorithm and evaluated its performance on two typical groups of surface samples. The performance of our new method is compared with both traditional ICP and weighted ICP. We also integrate a more elaborate scheme using normals for comparison. Our experiments show that WA-ICP has a good quality of convergence rate especially in the early iterations. The algorithm is also validated to be robust to low resolution and noise perturbation of 3D surfaces. Despite its simple form, the proposed WA-ICP clearly shows improved performance and some robust characteristics.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132316173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Script independent scene text segmentation using fast stroke width transform and GrabCut 脚本独立的场景文本分割使用快速笔画宽度变换和GrabCut
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486484
Jay H. Bosamiya, Palash Agrawal, P. Roy, B. Raman
{"title":"Script independent scene text segmentation using fast stroke width transform and GrabCut","authors":"Jay H. Bosamiya, Palash Agrawal, P. Roy, B. Raman","doi":"10.1109/ACPR.2015.7486484","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486484","url":null,"abstract":"In this paper, a novel script independent scene text segmentation method is presented. The proposed method introduces an approach for character segmentation which combines the advantages of connected component analysis, stroke width analysis and graph cuts. The method was evaluated on a standard dataset, and also on a custom dataset of scene images of text from various Indian scripts. Additionally, a fast implementation of Stroke Width Transform (SWT) is presented here, which improves speed over other implementations by 79%. From our experiments, we have achieved encouraging results on the usefulness of the combination of text/non-text classification with GrabCut. The technique is robust to the language/script of text.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114446634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Segmentation and recognition of text written in 3D using Leap motion interface 基于Leap运动界面的三维文本分割与识别
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486561
Chelsi Agarwal, D. P. Dogra, Rajkumar Saini, P. Roy
{"title":"Segmentation and recognition of text written in 3D using Leap motion interface","authors":"Chelsi Agarwal, D. P. Dogra, Rajkumar Saini, P. Roy","doi":"10.1109/ACPR.2015.7486561","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486561","url":null,"abstract":"In this paper, we present a word extraction and recognition methodology from online cursive handwritten text-lines recorded by Leap motion controller The online text, drawn by 3D gesture in air, is distinct from usual online pen-based strokes. The 3D gestures are recorded in air, hence they produce often non-uniform text style and jitter-effect while writing. Also, due to the constraint of writing in air, the pause of stroke-flow between words is missing. Instead all words and lines are connected by a continuous stroke. In this paper, we have used a simple but effective heuristic to segment words written in air. Here, we propose a segmentation methodology of continuous 3D strokes into text-lines and words. Separation of text lines is achieved by heuristically finding the large gap-information between end and start-positions of successive text lines. Word segmentation is characterized in our system as a two class problem. In the next phase, we have used Hidden Markov Model-based approach to recognize these segmented words. Our experimental validation with a large dataset consisting with 320 sentences reveals that the proposed heuristic based word segmentation algorithm performs with accuracy as high as 80.3%c and an accuracy of 77.6% has been recorded by HMM-based word recognition when these segmented words are fed to HMM. The results show that the framework is efficient even with cluttered gestures.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114719080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Principal affinity based cross-modal retrieval 基于主亲和力的跨模态检索
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486479
Jian Liang, Dong Cao, R. He, Zhenan Sun, T. Tan
{"title":"Principal affinity based cross-modal retrieval","authors":"Jian Liang, Dong Cao, R. He, Zhenan Sun, T. Tan","doi":"10.1109/ACPR.2015.7486479","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486479","url":null,"abstract":"Multimedia content is increasingly available in multiple modalities. Each modality provides a different representation of the same entity. This paper studies the problem of joint representation of the text and image components of multimedia documents. However, most existing algorithms focus more on inter-modal connection rather than intramodal feature extraction. In this paper, a simple yet effective principal affinity representation (PAR) approach is proposed to exploit the affinity representations of different modalities with local cluster samples. Afterwards, multi-class logistic regression model is adopted to learn the projections from principal affinity representation to semantic labels vectors. Inner product distance is further used to improve cross-modal retrieval performance. Extensive experiments on three benchmark datasets illustrate that our proposed method obtains significant improvements over the state-of-the-art subspace learning based cross-modal methods.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116065741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Multiple-dataset traffic sign classification with OneCNN 基于OneCNN的多数据集交通标志分类
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486576
Fran Jurisic, Ivan Filkovic, Z. Kalafatić
{"title":"Multiple-dataset traffic sign classification with OneCNN","authors":"Fran Jurisic, Ivan Filkovic, Z. Kalafatić","doi":"10.1109/ACPR.2015.7486576","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486576","url":null,"abstract":"We take a look at current state of traffic sign classification discussing what makes it a specific problem of visual object classification. With impressive state-of-the-art results it is easy to forget that the domain extends beyond annotated datasets and overlook the problems that must be faced before we can start training classifiers. We discuss such problems, give an overview of previous work done, go over publicly available datasets and present a new one. Following that, classification experiments are conducted using a single CNN model, deeper than used previously and trained with dropout. We apply it over multiple datasets from Germany, Belgium and Croatia, their intersections and union, outperforming humans and other single CNN architectures for traffic sign classification.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"163 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122135921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Skeleton based action recognition with convolutional neural network 基于骨架的卷积神经网络动作识别
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486569
Yong Du, Y. Fu, Liang Wang
{"title":"Skeleton based action recognition with convolutional neural network","authors":"Yong Du, Y. Fu, Liang Wang","doi":"10.1109/ACPR.2015.7486569","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486569","url":null,"abstract":"Temporal dynamics of postures over time is crucial for sequence-based action recognition. Human actions can be represented by the corresponding motions of articulated skeleton. Most of the existing approaches for skeleton based action recognition model the spatial-temporal evolution of actions based on hand-crafted features. As a kind of hierarchically adaptive filter banks, Convolutional Neural Network (CNN) performs well in representation learning. In this paper, we propose an end-to-end hierarchical architecture for skeleton based action recognition with CNN. Firstly, we represent a skeleton sequence as a matrix by concatenating the joint coordinates in each instant and arranging those vector representations in a chronological order. Then the matrix is quantified into an image and normalized to handle the variable-length problem. The final image is fed into a CNN model for feature extraction and recognition. For the specific structure of such images, the simple max-pooling plays an important role on spatial feature selection as well as temporal frequency adjustment, which can obtain more discriminative joint information for different actions and meanwhile address the variable-frequency problem. Experimental results demonstrate that our method achieves the state-of-art performance with high computational efficiency, especially surpassing the existing result by more than 15 percentage on the challenging ChaLearn gesture recognition dataset.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124869399","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 352
Semi-global depth from focus 半全局深度的焦点
2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2015-11-01 DOI: 10.1109/ACPR.2015.7486578
Wentao Liu, Xihong Wu
{"title":"Semi-global depth from focus","authors":"Wentao Liu, Xihong Wu","doi":"10.1109/ACPR.2015.7486578","DOIUrl":"https://doi.org/10.1109/ACPR.2015.7486578","url":null,"abstract":"Reconstructing depth map from a sequence of images with different camera settings, known as Depth from Focus (DFF), has been addressed for decades. However, Classical local methods, e.g., windowed filter, usually produce uncertain estimation in textureless areas and depth discontinuities. Meanwhile, they appear sensitive to filter size and image noise. In this research a semi-global depth from focus approach that considers both the accuracy and robustness is proposed to enforce an adaptive smoothness constraint based on modified Laplacian measure. Firstly an adaptive aggregation strategy is proposed to integrate the local focus measure and then an efficient optimization method, known as scanline optimization, is introduced to approximate the global optimal of depth estimation. To evaluate the effectiveness of this research, experiments on both synthetic and real datasets are performed. The experimental results show that the proposed approach generates more robust and accurate depth maps while preserving depth discontinuities.","PeriodicalId":240902,"journal":{"name":"2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129937604","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信