2018 IEEE Visual Communications and Image Processing (VCIP)最新文献

筛选
英文 中文
Compressed Sensing via a Deep Convolutional Auto-encoder 通过深度卷积自编码器压缩感知
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698640
Hao Wu, Ziyang Zheng, Yong Li, Wenrui Dai, H. Xiong
{"title":"Compressed Sensing via a Deep Convolutional Auto-encoder","authors":"Hao Wu, Ziyang Zheng, Yong Li, Wenrui Dai, H. Xiong","doi":"10.1109/VCIP.2018.8698640","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698640","url":null,"abstract":"The nonlinear recovery is not promising in accuracy and speed, which limits the practical usage of compressed sensing (CS). This paper proposes a deep learning-based CS framework which leverages a deep convolutional auto-encoder for image sensing and recovery. The utilized auto-encoder architecture consists of three components: the fully convolutional network acts as an adaptive measurement matrix generator in the encoder; while in the decoder, the deconvolution network and refined reconstruction network are learned for intermediate and final recovery, respectively. Different from most previous work focusing on the block-wise manner to reduce implementation cost but result in blocky artifacts, our adaptive measurement matrix is applicable to any size of scene image and the decoder network reconstructs the whole image efficiently without any blocky artifacts. Moreover, dense connectivity is leveraged to combine multi-level features and alleviate the vanishing-gradient problem in the refined reconstruction network which boosts the performance on image recovery. Compared to the state-of-the-art methods, our algorithm improves more than 0.8 dB in average PSNR.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133964233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Voting-based Hand-Waving Gesture Spotting from a Low-Resolution Far-Infrared Image Sequence 低分辨率远红外图像序列中基于投票的手势识别
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698650
Yasutomo Kawanishi, Chisato Toriyama, Tomokazu Takahashi, Daisuke Deguchi, I. Ide, H. Murase, Tomoyoshi Aizawa, M. Kawade
{"title":"Voting-based Hand-Waving Gesture Spotting from a Low-Resolution Far-Infrared Image Sequence","authors":"Yasutomo Kawanishi, Chisato Toriyama, Tomokazu Takahashi, Daisuke Deguchi, I. Ide, H. Murase, Tomoyoshi Aizawa, M. Kawade","doi":"10.1109/VCIP.2018.8698650","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698650","url":null,"abstract":"We propose a temporal spotting method of a hand gesture from a low-resolution far-infrared image sequence captured by a far-infrared sensor array. The sensor array captures the spatial distribution of far-infrared intensity as a thermal image by detecting far-infrared waves emitted from heat sources. It is difficult to spot a hand gesture from a sequence of thermal images captured by the sensor due to its low-resolution, heavy noise, and varying duration of the gesture. Therefore, we introduce a voting-based approach to spot the gesture with template matching-based gesture recognition. We confirm the effectiveness of the proposed temporal spotting method in several settings.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132577591","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Fast Korean Text Detection and Recognition in Traffic Guide Signs 交通引导标志中朝鲜语文本的快速检测与识别
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698668
Hyunjun Eun, Jonghee Kim, Jinsu Kim, Changick Kim
{"title":"Fast Korean Text Detection and Recognition in Traffic Guide Signs","authors":"Hyunjun Eun, Jonghee Kim, Jinsu Kim, Changick Kim","doi":"10.1109/VCIP.2018.8698668","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698668","url":null,"abstract":"In this paper, we propose a fast method based on deep neural networks to detect and recognize Korean characters in traffic guide signs. To detect character candidates quickly, we first employ a region proposal network (RPN) which is in this paper ResNet-18, being relatively shallow. We also apply the Inception architecture to residual blocks for reducing parameters of the network. After character candidates are detected, we classify them into 709 Korean characters by using a classification network (CLSN). Similar to the RPN, our CLSN consists of residual blocks with the Inception architecture. In experiments, we achieved 97.69 % of accuracy at 5.9fps on both detection and recognition of Korean characters in traffic guide signs.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"134 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123210616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Light Field Image Sparse Coding via CNN-Based EPI Super-Resolution 基于cnn的EPI超分辨率光场图像稀疏编码
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698714
Jinbo Zhao, P. An, Xinpeng Huang, Liang Shan, Ran Ma
{"title":"Light Field Image Sparse Coding via CNN-Based EPI Super-Resolution","authors":"Jinbo Zhao, P. An, Xinpeng Huang, Liang Shan, Ran Ma","doi":"10.1109/VCIP.2018.8698714","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698714","url":null,"abstract":"This paper proposes a novel light field (LF) image compression scheme by super resolving the epipolar plane image (EPI) via convolutional neural network (CNN). In the scheme, we first decompose the LF image into sub-aperture images (SAIs), and only one quarter of them are compressed on the encoding side to reduce the bitrate. On the decoding side, we use these selected SAIs to reconstruct the entire LF by taking advantage of the special structure of EPI. The low-resolution EPIs generated from the sparse SAIs are super resolved by using deep residual network and the output high-resolution EPIs are used to rebuild the dense SAIs. Experimental results show the superior performance of our scheme, which achieve 1.46 dB quality improvement and 35.85 percent bit rate reduction on average compared with the typical pseudo-sequence-based coding method.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"33 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123218380","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Near-Duplicate Image Retrieval Based on Multiple Features 基于多特征的近重复图像检索
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698664
Xueqin Zhang
{"title":"Near-Duplicate Image Retrieval Based on Multiple Features","authors":"Xueqin Zhang","doi":"10.1109/VCIP.2018.8698664","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698664","url":null,"abstract":"In this paper, we propose a near-duplicate image retrieval method based on multiple features. Combining the deep features extracted from the VGG relu6 layer with the improved local feature descriptors, we attempt to simulate the near-duplicate image retrieval process of the human brain through a two-layer retrieval structure. Inspired by the proposed CROW feature, we calculate the weights on VGG shallow pooling layer and extract the interest domains for screening surf feature points. At the same time, a center weight is proposed to improve the VLAD algorithm. Experiments show that our method can not only obtain the visually similar results of an image, but also obtain the results that contain the visually prominent parts of the image.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115842609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rate-Distortion Theory for Simplified Affine Motion Compensation Used in Video Coding 用于视频编码的简化仿射运动补偿的率失真理论
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698702
H. Meuel, Stephan Ferenz, Yiqun Liu, J. Ostermann
{"title":"Rate-Distortion Theory for Simplified Affine Motion Compensation Used in Video Coding","authors":"H. Meuel, Stephan Ferenz, Yiqun Liu, J. Ostermann","doi":"10.1109/VCIP.2018.8698702","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698702","url":null,"abstract":"In this work, we derive the rate-distortion function for video coding using the simplified affine, 4-parameter motion compensation model as it is used in the Joint Exploration Model (JEM) by the Joint Video Exploration Team (JVET) on Future Video coding. We model the displacement estimation error during motion estimation and obtain the bit rate by applying the rate-distortion theory. We assume that the displacement estimation error is caused by perturbed parameters of the simplified affine model. These transformation parameters are assumed statistically independent, with each of them having a zero-mean Gaussian distributed estimation error. The joint probability density function (p.d.f.) of the displacement estimation errors is derived and related to the prediction error. We calculate the bit rate as a function of the accuracy of the parameter estimation for the simplified affine motion model. Finally, we compare our results with a translational motion model as used in video coding standards like HEVC as well as with a full affine motion model with 6 degrees of freedom. For aerial sequences containing distinct affine motion, the minimum required bit rate to encode the prediction error can be significantly reduced from 2.5 bit/sample to 0.02 bit/sample for a reasonable operating point and a block size of 64×64 pel2.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116656038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Simple Iterative Clustering on Graphs for Robust Model Fitting 图上简单迭代聚类的鲁棒模型拟合
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698736
H. Luo, Guobao Xiao, Hanzi Wang
{"title":"Simple Iterative Clustering on Graphs for Robust Model Fitting","authors":"H. Luo, Guobao Xiao, Hanzi Wang","doi":"10.1109/VCIP.2018.8698736","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698736","url":null,"abstract":"In this paper, we propose a novel method, simple iterative clustering on graphs (SICG), to deal with robust model fitting problems. Specifically, we first construct a graph, where each vertex denotes a model hypothesis and each edge represents the similarity between two model hypotheses, for model fitting. We then propose a simple iterative clustering algorithm, which adapts the k-medoids clustering algorithm, to intuitively estimate model instances in data. The proposed SICG method is able to effectively fit and segment multiple-structure data contaminated with a large number of outliers and noises. Experimental results show that SICG achieves superior fitting results over several state-of-the-art model fitting methods on real images.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130391504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Optimized Spatial Recurrent Network for Intra Prediction in Video Coding 基于优化空间循环网络的视频编码内预测
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698658
Yueyu Hu, Wenhan Yang, Sifeng Xia, Jiaying Liu
{"title":"Optimized Spatial Recurrent Network for Intra Prediction in Video Coding","authors":"Yueyu Hu, Wenhan Yang, Sifeng Xia, Jiaying Liu","doi":"10.1109/VCIP.2018.8698658","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698658","url":null,"abstract":"Intra prediction in modern video codecs is able to efficiently reduce spatial redundancy in video frames. With preceding pixels as context, traditional intra prediction schemes generate linear predictions based on several predefined directions (i.e. modes) for the current prediction unit (PU). However, these modes are relatively simple and are not able to handle complex textures, which leads to additional bits encoding the residue. In this paper, we design a convolutional neural network (CNN) guided spatial recurrent neural network (RNN) to improve the intra prediction in High-Efficiency Video Coding (HEVC). By exploring the correlations between pixels, the network learns to generate prediction signal in a progressive manner. The progressive model solves the problem of asymmetry in intra prediction naturally. As the model is designed for global context modeling, no flags for intra prediction modes selection need to be encoded. Our proposed intra prediction scheme achieves on average 1.2% bit-rate saving compared with HEVC.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134076521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Advanced Orientation Robust Face Detection Algorithm Using Prominent Features and Hybrid Learning Techniques 基于突出特征和混合学习技术的高级定向鲁棒人脸检测算法
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698649
Chien-Yu Chen, Jian-Jiun Ding, H. Hsu, Yih-Cherng Lee
{"title":"Advanced Orientation Robust Face Detection Algorithm Using Prominent Features and Hybrid Learning Techniques","authors":"Chien-Yu Chen, Jian-Jiun Ding, H. Hsu, Yih-Cherng Lee","doi":"10.1109/VCIP.2018.8698649","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698649","url":null,"abstract":"Face detection is one of the most popular topics in computer vision. There are several well-known techniques for face detection, such as the Viola-Jones detector. However, the performance of the Viola-Jones detector is limited since it mainly applies the simple Haar-based features. Many advanced methods, especially the convolutional neural network (CNN) based method, have very good performance in face detection. However, they require huge amount of training data. Moreover, most of existing algorithms are not robust to rotation, head-up, and head-down cases. In this paper, we find that, with some modifications, the Viola-Jones detector can also have very good performance in face detection. In addition to the Haar features, we also apply the prominent features and the color information. With the contour information, the edge-aware filter, the background smoother, the fuzzy classifier, and the relative locations, the prominent features, such as eyes, mouths, noses, and ears, can be extracted accurately. With these features, the accuracy of face detection can be much improved. Simulations show that, even if huge amount of training data is not applied, the proposed algorithm has better performance than state-of-the-art face detection methods, including the CNN-based method.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"315 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132553615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Weighted Two-Phase Linear Reconstruction Measure-based Classification 基于加权两相线性重构测度的分类方法
2018 IEEE Visual Communications and Image Processing (VCIP) Pub Date : 2018-12-01 DOI: 10.1109/VCIP.2018.8698656
Jianping Gou, Jun Song, Heping Song, Liangjun Wang
{"title":"Weighted Two-Phase Linear Reconstruction Measure-based Classification","authors":"Jianping Gou, Jun Song, Heping Song, Liangjun Wang","doi":"10.1109/VCIP.2018.8698656","DOIUrl":"https://doi.org/10.1109/VCIP.2018.8698656","url":null,"abstract":"Linear reconstruction measure (LRM) is a promising similarity measure of data. In this paper, we consider the locality of data in LRM, and propose weighted two-phase linear reconstruction measure-based classification (WTPLRMC). In WTPLRMC, the first phase determines the representative training samples from all training samples by LRM, and the second phase constrains the linear reconstruction coefficients of the chosen representative training samples in first phase using the locality of data, which is reflected by the similarity weights between each test sample and the representative training samples. The effectiveness of the proposed WTPLRMC is well demonstrated on some benchmark face databases with satisfactory classification results.","PeriodicalId":270457,"journal":{"name":"2018 IEEE Visual Communications and Image Processing (VCIP)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131851074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信