2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)最新文献

筛选
英文 中文
Fast Genre Classification of Web Images Using Global and Local Features 使用全局和局部特征的Web图像快速类型分类
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.84
Guo-Shuai Liu, Fei Yin, Zhenbo Luo, Cheng-Lin Liu
{"title":"Fast Genre Classification of Web Images Using Global and Local Features","authors":"Guo-Shuai Liu, Fei Yin, Zhenbo Luo, Cheng-Lin Liu","doi":"10.1109/ACPR.2017.84","DOIUrl":"https://doi.org/10.1109/ACPR.2017.84","url":null,"abstract":"A number of images are present on the Web and the number is increasing every day. To effectively mine the contents embedded in Web images, it is useful to classify the images into different types so that they can be fed to different procedures for detailed analysis, such as text and non-text image discrimination. We herein propose a hierarchical algorithm for efficiently classifying Web images into four classes, namely, natural scene images, born-digital images, scanned and cameracaptured paper documents, which are the most prevalent image types on the Web. Our algorithm consists of two stages; the first stage extracts global features reflecting the distributions of color, edge and gradient, and uses a support vector machine (SVM) classifier for preliminary classification. Images assigned low confidence by the first-stage classifier is processed by the second stage, which further extracts local texture features represented in the Bag-of-Words framework and uses another SVM classifier for final classification. In addition, we design two fusion strategies to train the second classifier and generate the final prediction label depending on the usage of local features in the second stage. To validate the effectiveness of our proposed method, we also build a database containing more than 55,000 images from various sources. On our test image set, we obtained an overall classification accuracy of 98.4% and the processing speed is over 27FPS on an Intel(R) Xeon(R) CPU (2.90GHz).","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122338153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Single Image Super-Resolution via Mixed Examples and Sparse Representation 基于混合样例和稀疏表示的单幅图像超分辨率
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.110
Weirong Liu, Changhong Shi, Chaorong Liu, Jie Liu
{"title":"Single Image Super-Resolution via Mixed Examples and Sparse Representation","authors":"Weirong Liu, Changhong Shi, Chaorong Liu, Jie Liu","doi":"10.1109/ACPR.2017.110","DOIUrl":"https://doi.org/10.1109/ACPR.2017.110","url":null,"abstract":"Existing super-resolution (SR) methods can be divided into two classes: the external examples SR and the internal examples SR. Although these two types of methods have been achieved satisfactory results, such methods are limited by their inherent flaws. This paper proposes mixed example selection method for combining the external examples with the internal examples. We cluster the internal examples into K classes, and select the similar external examples for every cluster to enrich the training database. And then we learn K discriminative dictionaries for the K cluster examples. Finally, we reconstruct the low resolution images with the learned discriminative dictionaries. Experiments validate the effectiveness of the proposed method in terms of visual and quantitative assessments.","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123751143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enlarging Effective Receptive Field of Convolutional Neural Networks for Better Semantic Segmentation 扩大卷积神经网络的有效接受野以实现更好的语义分割
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.7
Yifan Gu, Zuofeng Zhong, Shuai Wu, Yong Xu
{"title":"Enlarging Effective Receptive Field of Convolutional Neural Networks for Better Semantic Segmentation","authors":"Yifan Gu, Zuofeng Zhong, Shuai Wu, Yong Xu","doi":"10.1109/ACPR.2017.7","DOIUrl":"https://doi.org/10.1109/ACPR.2017.7","url":null,"abstract":"Recently, convolutional neural networks have shown powerful capability in different fields of computer vision, and have become the most effective means for dense prediction problems such as semantic segmentation. However, methods based on fully convolution network(FCN) are inherently limited to the size of the receptive field for each pixel, which leads to the bad performance of predicting object boundary. In this paper, we propose a novel deep neural network module, namely group dilated convolution(GDC), to effectively enlarge the receptive field, and a top-to-down pathway network is exploited simultaneously. The idea is that dilation convolution with different ratios can cover features of different scales, which shows a significant Mean IOU improvement in comparison with the baseline network.","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128690661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Hierarchical Classification Strategy for Robust Detection of Passive/Active Mental State Using User-Voluntary Pitch Imagery Task 基于用户自主音高意象任务的被动/主动心理状态鲁棒检测层次分类策略
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.133
Young-Jin Kee, Min-Ho Lee, J. Williamson, Seong-Whan Lee
{"title":"A Hierarchical Classification Strategy for Robust Detection of Passive/Active Mental State Using User-Voluntary Pitch Imagery Task","authors":"Young-Jin Kee, Min-Ho Lee, J. Williamson, Seong-Whan Lee","doi":"10.1109/ACPR.2017.133","DOIUrl":"https://doi.org/10.1109/ACPR.2017.133","url":null,"abstract":"Event-related potentials (ERPs) represent neuronal activity in the brain elicited by external visual/auditory stimulation, and it is widely used in brain-computer interface (BCI) systems. The ERP responses are elicited a few milliseconds after attending to an oddball stimulus; target and non-target stimulus are repeatedly flashed while the electroencephalography (EEG) is recording. ERP responses in the EEG signal have a poor signal-to-ratio in single-trial analysis; therefore, the epochs of the target and non-target trials are averaged over time in order to improve their decoding accuracy. Furthermore, these exogenous potentials can be naturally evoked by just looking at a target symbol. Therefore, the BCI system could generate unintended commands without considering the user's intention. In this study, we approach this dilemma by assuming that a greater effort for the mental task would evoke a stronger positive/negative ERP deflection. Three mental states are defined: passive gazing, active counting, and pitch-imagery. The experiments results showed significantly enhanced ERP patterns and averaged decoding accuracies of 80%, 95.4%, and 95.6%, respectively. The decoding accuracies between both active tasks and the passive task showed an averaged accuracy of 57.5% (gazing vs. counting) and 72.5% (gazing vs. pitch-imagery). Following this result, we proposed a hierarchy classification strategy where the passive or active mental state is decoded in the first stage, and the target stimuli are estimated in the second stage. Our work is the first to propose a system that classifies an intended or unintended brain state by considering the measurable differences of mental effort in the EEG signal so that unintended commands to the system are minimized.","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116744420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Complete Dual-Cross Pattern for Unconstrained Texture Classification 一种用于无约束纹理分类的完全双交叉模式
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.160
S. K. Roy, B. Chanda, B. Chaudhuri, D. Ghosh, S. Dubey
{"title":"A Complete Dual-Cross Pattern for Unconstrained Texture Classification","authors":"S. K. Roy, B. Chanda, B. Chaudhuri, D. Ghosh, S. Dubey","doi":"10.1109/ACPR.2017.160","DOIUrl":"https://doi.org/10.1109/ACPR.2017.160","url":null,"abstract":"In order to perform unconstrained texture classification, this paper presents a novel and computationally efficient texture descriptor called Complete Dual-Cross Pattern (CDCP), which is robust to gray-scale changes and surface rotation. To extract CDCP, at first a gray scale normalization scheme is used to reduce the illumination effect and, then CDCP feature is computed from holistic and component levels. A local region of the texture image is represented by it's center pixel and difference of sign-magnitude transform (DSMT) at multiple levels. Using a global threshold, the gray value of center pixel is converted into a binary code named DCP center (DCP_C). DSMT decomposes into two complementary components: the sign and the magnitude. They are encoded respectively into DCP-sign (DCP_S) and DCP-magnitude (DCP_M), based on their corresponding threshold values. Finally, CDCP is formed by fusing DCP_S, DCP_M and DCP_C features through joint distribution. The invariance characteristics of CDCP are attained due to computation of pattern at multiple levels, which makes CDCP highly discriminative and achieves state-of-the-art performance for rotation invariant texture classification.","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"155 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114297026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Online Optimization and Feedback Elman Neural Network for Maneuvering Target Tracking 机动目标跟踪的在线优化与反馈Elman神经网络
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.56
L. Xia, Ya Zhang, Huajun Liu
{"title":"Online Optimization and Feedback Elman Neural Network for Maneuvering Target Tracking","authors":"L. Xia, Ya Zhang, Huajun Liu","doi":"10.1109/ACPR.2017.56","DOIUrl":"https://doi.org/10.1109/ACPR.2017.56","url":null,"abstract":"The uncertainty of maneuver model and nonlinear filtering, which are two difficult problems in practical application of maneuvering target tracking, are becoming the focus of research. Based on this, we propose an online maneuvering target tracking filter algorithm based on Elman neural network which can feedback while optimizing the estimation. Based on the Constant Acceleration (CA) model, the Elman neural network algorithm is used to obtain the size of the target maneuver and adaptive adjustment factor of noise covariance matrix, by online learning of the difference of the target state prediction and the optimal estimation, the innovation and the filter gain matrix, to real-time adjust optimal estimation and motion model. Mass of simulation experiments show that the proposed algorithm can effectively reduce the interference of the maneuvering of targets to the motion model during the target motion and improve the filtering performance. Under the condition of strong maneuvering, the tracking performance is far superior to Singer model, and also better than the IMM_ELM tracking filter algorithm.","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126771248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Probability Based Voting for Vanishing Point Detection 基于概率投票的消失点检测
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.39
Qian Chen, Kouki Masumoto, Haiyuan Wu, S. Lao
{"title":"Probability Based Voting for Vanishing Point Detection","authors":"Qian Chen, Kouki Masumoto, Haiyuan Wu, S. Lao","doi":"10.1109/ACPR.2017.39","DOIUrl":"https://doi.org/10.1109/ACPR.2017.39","url":null,"abstract":"This paper describes a method to detect vanishing points based on voting with probability. We use a group of lines in stead of a single line to describe the line passes a set of feature points belonging to the same line segment and propose a probability distribution to describe the group of lines. We use it to estimate the probability of all the points on the image plane that can be considered on the detected line. We use the surface of a cube centered at the projection center as the space for voting. We accumulate the probability of each point on the image plane (including the points at infinity) calculated with all the detected lines and project it on to the cube surface. The vanishing points are then detected by finding the cells that are local maximum of the accumulated probabilities and have many lines passed them. Through the comparative experiments, we confirmed that our method could give accurate and stable results and our method is not sensitive to the resolution of the voting space.","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"175 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127630722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fully Convolutional DenseNet for Saliency-Map Prediction 用于显著性图预测的全卷积密度网络
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.143
Taiki Oyama, Takao Yamanaka
{"title":"Fully Convolutional DenseNet for Saliency-Map Prediction","authors":"Taiki Oyama, Takao Yamanaka","doi":"10.1109/ACPR.2017.143","DOIUrl":"https://doi.org/10.1109/ACPR.2017.143","url":null,"abstract":"In this paper, we propose a fully convolutional DenseNet model for saliency-map prediction (DenseSal). While the most state-of-the-art models for predicting saliency maps use shallow networks such as VGG-16, our model uses densely connected convolutional networks (DenseNet) with over 150 layers. Since DenseNet has shown the excellent results on image classification tasks, the coarse feature maps from the fully convolutional neural networks based on DenseNets were concatenated to predict saliency maps through a readout network. It is shown that the DenseNet is useful for the saliency-map prediction and achieved the state-of-the-art accuracy on the major fixation datasets.","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132332654","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Robust Multi-scale ORB Algorithm in Real-Time Monocular Visual Odometry 实时单目视觉里程测量中的鲁棒多尺度ORB算法
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.101
Qiongjie Cui, Huajun Liu
{"title":"Robust Multi-scale ORB Algorithm in Real-Time Monocular Visual Odometry","authors":"Qiongjie Cui, Huajun Liu","doi":"10.1109/ACPR.2017.101","DOIUrl":"https://doi.org/10.1109/ACPR.2017.101","url":null,"abstract":"In this paper, a novel multi-scale ORB algorithm with lower computation is proposed applied for increasing correct matches of feature points in visual odometry when image scale changes. Since ORB algorithm has little scale invariance for feature points matching, the visual odometry employing the ORB algorithm directly performs poorly in the position and orientation estimation. Therefore, the proposed algorithm combines the ORB with SURF by added the scale space. In addition, single layer non-maximum suppression is applied to the selection of stable feature points to decrease spending time in matching step. Experimental results present that the proposed algorithm achieves good matching performance in terms with scale invariance taking into consideration. It was found that the position estimation and the orientation estimation was improved compared to the visual odometry based on the ORB algorithm while the spend of time has only increased a little.","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130378761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-feature Joint Dictionary Learning for Face Recognition 人脸识别的多特征联合字典学习
2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) Pub Date : 2017-11-01 DOI: 10.1109/ACPR.2017.138
Meng Yang, Qiangchang Wang, Wei Wen, Zhihui Lai
{"title":"Multi-feature Joint Dictionary Learning for Face Recognition","authors":"Meng Yang, Qiangchang Wang, Wei Wen, Zhihui Lai","doi":"10.1109/ACPR.2017.138","DOIUrl":"https://doi.org/10.1109/ACPR.2017.138","url":null,"abstract":"Dictionary learning with sparse representation has been widely used for pattern classification tasks, where an input is classified to the category with the minimum reconstruction error. While most methods focus on singlefeature recognition problems, recent studies have proved the superiorities of exploiting multi-feature fusion classification. In this paper, we present a new multi-feature joint dictionary learning algorithm which can enhance correlations among different features via our designed classlevel similarity regularization. The proposed algorithm can fuse different information and correlate these dictionary atoms within the same pattern category. Besides, the distinctiveness of several features is weighted differently to reflect their discriminative abilities. Furthermore, a dictionary learning algorithm is used to reduce dictionary size. The proposed algorithm achieves comparable experimental results in several face recognition databases.","PeriodicalId":426561,"journal":{"name":"2017 4th IAPR Asian Conference on Pattern Recognition (ACPR)","volume":"232 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131994681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信