2014 IEEE Visual Communications and Image Processing Conference最新文献

筛选
英文 中文
Disocclusion hole-filling in DIBR-synthesized images using multi-scale template matching 基于多尺度模板匹配的dibr合成图像去咬合补孔
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-07 DOI: 10.1109/VCIP.2014.7051614
S. Reel, Kam Cheung Patrick Wong, Gene Cheung, L. Dooley
{"title":"Disocclusion hole-filling in DIBR-synthesized images using multi-scale template matching","authors":"S. Reel, Kam Cheung Patrick Wong, Gene Cheung, L. Dooley","doi":"10.1109/VCIP.2014.7051614","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051614","url":null,"abstract":"Transmitting texture and depth images of captured camera view(s) of a 3D scene enables a receiver to synthesize novel virtual viewpoint images via Depth-Image-Based Rendering (DIBR). However, a DIBR-synthesized image often contains disocclusion holes, which are spatial regions in the virtual view image that were occluded by foreground objects in the captured camera view(s). In this paper, we propose to complete these disocclusion holes by exploiting the self-similarity characteristic of natural images via nonlocal template-matching (TM). Specifically, we first define self-similarity as nonlocal recurrences of pixel patches within the same image across different scales-one characterization of self-similarity in a given image is the scale range in which these patch recurrences take place. Then, at encoder we segment an image into multiple depth layers using available per-pixel depth values, and characterize self-similarity in each layer with a scale range; scale ranges for all layers are transmitted as side information to the decoder. At decoder, disocclusion holes are completed via TM on a per-layer basis by searching for similar patches within the designated scale range. Experimental results show that our method improves the quality of rendered images over previous disocclusion hole-filling algorithms by up to 3.9dB in PSNR.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"120 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116636554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Robust image registration using adaptive expectation maximisation based PCA 基于PCA的自适应期望最大化鲁棒图像配准
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-07 DOI: 10.1109/VCIP.2014.7051515
P. Reel, L. Dooley, Kam Cheung Patrick Wong, A. Börner
{"title":"Robust image registration using adaptive expectation maximisation based PCA","authors":"P. Reel, L. Dooley, Kam Cheung Patrick Wong, A. Börner","doi":"10.1109/VCIP.2014.7051515","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051515","url":null,"abstract":"Images having either the same or different modalities can be aligned using the systematic process of image registration. Inherent image characteristics including intensity non-uniformities in magnetic resonance images and large homogeneous non-vascular regions in retinal and other generic image types however, pose a significant challenge to their registration. This paper presents an adaptive expectation maximisation for principal component analysis with mutual information (aEMPCA-MI) similarity measure for image registration. It introduces a novel iterative process to adaptively select the most significant principal components using Kaiser rule and applies 4-pixel connectivity for feature extraction together with Wichard's bin size selection in calculating the MI. Both quantitative and qualitative results on a diverse range of image datasets, conclusively demonstrate the superior image registration performance of aEMPCA-MI compared with existing Mi-based similarity measures.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128406367","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Rate-distortion optimised transform competition for intra coding in HEVC HEVC中率失真优化的帧内编码变换竞争
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-07 DOI: 10.1109/VCIP.2014.7051507
A. Arrufat, P. Philippe, O. Déforges
{"title":"Rate-distortion optimised transform competition for intra coding in HEVC","authors":"A. Arrufat, P. Philippe, O. Déforges","doi":"10.1109/VCIP.2014.7051507","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051507","url":null,"abstract":"State of the art video coders are based on prediction and transform coding. The transform decorrelates the signal to achieve high compression levels. In this paper we propose improving the performances of the latest video coding standard, HEVC, by adding a set of rate-distortion optimised transforms (RDOTs). The transform design is based upon a cost function that incorporates a bit rate constraint. These new RDOTs compete against classical HEVC transforms in the rate-distortion optimisation (RDO) loop in the same way as prediction modes and block sizes, providing additional coding possibilities. Reductions in BD-rate of around 2% are demonstrated when making these transforms available in HEVC.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125613878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A joint 3D image semantic segmentation and scalable coding scheme with ROI approach 一种结合ROI方法的三维图像语义分割和可扩展编码方案
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-07 DOI: 10.1109/VCIP.2014.7051556
Khouloud Samrouth, O. Déforges, Yi Liu, W. Falou, Mohamad Khalil
{"title":"A joint 3D image semantic segmentation and scalable coding scheme with ROI approach","authors":"Khouloud Samrouth, O. Déforges, Yi Liu, W. Falou, Mohamad Khalil","doi":"10.1109/VCIP.2014.7051556","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051556","url":null,"abstract":"Along with the digital evolution, image post-production and indexing have become one of the most advanced and desired services in the lossless 3D image domain. The 3D context provides a significant gain in terms of semantics for scene representation. However, it also induces many drawbacks including monitoring visual degradation of compressed 3D image (especially upon edges), and increased complexity for scene representation. In this paper, we propose a semantic region representation and a scalable coding scheme. First, the semantic region representation scheme is based on a low resolution version of the 3D image. It provides the possibility to segment the image according to a desirable balance between 2D and depth. Second, the scalable coding scheme consists in selecting a number of regions as a Region of Interest (RoI), based on the region representation, in order to be refined at a higher bitrate. Experiments show that the proposed scheme provides a high coherence between texture, depth and regions and ensures an efficient solution to the problems of compression and scene representation in the 3D image domain.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"11 7","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114002988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Non-separable mode dependent transforms for intra coding in HEVC HEVC中不可分离模式相关的帧内编码变换
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-07 DOI: 10.1109/VCIP.2014.7051504
A. Arrufat, P. Philippe, O. Déforges
{"title":"Non-separable mode dependent transforms for intra coding in HEVC","authors":"A. Arrufat, P. Philippe, O. Déforges","doi":"10.1109/VCIP.2014.7051504","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051504","url":null,"abstract":"Transform coding plays a crucial role in video coders. Recently, additional transforms based on the DST and the DCT have been included in the latest video coding standard, HEVC. Those transforms were introduced after a thoroughly analysis of the video signal properties. In this paper, we design additional transforms by using an alternative learning approach. The appropriateness of the design over the classical KLT learning is also shown. Subsequently, the additional designed transforms are applied to the latest HEVC scheme. Results show that coding performance is improved compared to the standard. Additional results show that the coding performance can be significantly further improved by using non-separable transforms. Bitrate reductions in the range of 2% over HEVC are achieved with those proposed transforms.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131922137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Optimized spatial and temporal resolution based on subjective quality estimation without encoding 基于主观质量估计的优化时空分辨率,无需编码
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051497
M. Takagi, H. Fujii, A. Shimizu
{"title":"Optimized spatial and temporal resolution based on subjective quality estimation without encoding","authors":"M. Takagi, H. Fujii, A. Shimizu","doi":"10.1109/VCIP.2014.7051497","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051497","url":null,"abstract":"In this paper, we propose a method of estimating subjective video quality with various spatial and temporal resolutions without encoding. Under a given bitrate constraint, the combination of resolution and frame rate that provides best subjective video quality depends on the video content. To maximize subjective video quality, several studies have proposed models that can estimate subjective quality with various resolutions and frame rates. However, to determine the optimal resolution and frame rate that maximize subjective video quality, it is necessary to estimate subjective video quality at each combination of resolution/frame rate/bitrate. This takes considerable time with previously reported methods because they require an encoding process for decoding videos or obtaining pre analysis. To address this issue, we developed a method that does not require an encoding process to estimate subjective video quality.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116958530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Complexity control of HEVC based on region-of-interest attention model 基于兴趣区域注意模型的HEVC复杂度控制
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051545
Xin Deng, Mai Xu, Shengxi Li, Zulin Wang
{"title":"Complexity control of HEVC based on region-of-interest attention model","authors":"Xin Deng, Mai Xu, Shengxi Li, Zulin Wang","doi":"10.1109/VCIP.2014.7051545","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051545","url":null,"abstract":"In this paper, we present a novel complexity control method of HEVC to adjust its encoding complexity. First, a region-of-interest (ROI) attention model is established, which defines different weights for various regions according to their importance. Then, the complexity control algorithm is proposed with a distortion-complexity optimization model, to determine the maximum depth of the largest coding units (LCUs) according to their weights. We can reduce the encoding complexity to a given target level at the cost of little distortion loss. Finally, the experimental results show that the encoding complexity can drop to a pre-defined target complexity as low as 20% with bias less than 7%. Meanwhile, our method is verified to preserve the quality of ROI better than another state-of-the-art approach.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"418 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127076880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Accelerated hybrid image reconstruction for non-regular sampling color sensors 非规则采样颜色传感器的加速混合图像重建
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051543
M. Bätz, Andrea Eichenseer, Markus Jonscher, Jürgen Seiler, André Kaup
{"title":"Accelerated hybrid image reconstruction for non-regular sampling color sensors","authors":"M. Bätz, Andrea Eichenseer, Markus Jonscher, Jürgen Seiler, André Kaup","doi":"10.1109/VCIP.2014.7051543","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051543","url":null,"abstract":"Increasing the spatial resolution is an ongoing research topic in image processing. A recently presented approach applies a non-regular sampling mask on a low resolution sensor and subsequently reconstructs the masked area via an extrapolation algorithm to obtain a high resolution image. This paper introduces an acceleration of this approach for use with full color sensors. Instead of employing the effective, yet computationally expensive extrapolation algorithm on each of the three RGB channels, a color space conversion is performed and only the luminance channel is then reconstructed using this algorithm. As natural images contain much less information in the chrominance channels, a fast linear interpolation technique can here be used to accelerate the whole reconstruction procedure. Simulation results show that an average speed up factor of 2.9 is thus achieved, while the loss in visual quality stays imperceptible. Comparisons of PSNR results confirm this.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126196543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A hardware-oriented IME algorithm and its implementation for HEVC 一种面向硬件的HEVC IME算法及其实现
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051540
Xin Ye, Dandan Ding, Lu Yu
{"title":"A hardware-oriented IME algorithm and its implementation for HEVC","authors":"Xin Ye, Dandan Ding, Lu Yu","doi":"10.1109/VCIP.2014.7051540","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051540","url":null,"abstract":"The flexible coding structure in High Efficiency Video Coding (HEVC) introduces many challenges to real-time implementation of the integer-pel motion estimation (IME). In this paper, a hardware-oriented IME algorithm naming parallel clustering tree search (PCTS) is proposed, where various prediction units (PU) are processed simultaneously with a parallel scheme. The PCTS consists of four hierarchical search steps. After each search step, PUs with the same MV candidate are clustered to one group. And the next search step is shared by PUs in the same group. Owing to the top-down tree-structure search strategy of the PCTS, search processes are highly shared among different PUs and system throughput is thus significantly increased. As a result, the hardware implementation based on the proposed algorithm can support real-time video applications of QFHD (3840×2160) at 30fps.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125390856","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Fast mode decision method for all intra spatial scalability in SHVC 基于全空间可扩展性的SHVC快速模式决策方法
2014 IEEE Visual Communications and Image Processing Conference Pub Date : 2014-12-01 DOI: 10.1109/VCIP.2014.7051589
Xuguang Zuo, Lu Yu
{"title":"Fast mode decision method for all intra spatial scalability in SHVC","authors":"Xuguang Zuo, Lu Yu","doi":"10.1109/VCIP.2014.7051589","DOIUrl":"https://doi.org/10.1109/VCIP.2014.7051589","url":null,"abstract":"Scalable high efficiency video coding (SHVC) is now being developed by the Joint Collaborative Team on Video Coding (JCT-VC). In SHVC, the enhancement layer (EL) employs the same tree structured coding unit (CU) and 35 intra prediction modes as the base layer (BL), which results in heavy computation load. To speed up the mode decision process in the EL, the correlations of the CU depth and intra prediction modes between the BL and the EL are exploited in this paper. Based on the correlations an EL CU depth early skip algorithm and a fast intra prediction mode decision algorithm are proposed for all intra spatial scalability. Experimental results show that 45.3% and 42.3% coding time of the EL can be saved in AH Intra 1.5× spatial scalability and 2× spatial scalability respectively. In the meantime, the R-D performance degraded less than 0.05% compared with SHVC Test Model (SHM) 5.0.","PeriodicalId":166978,"journal":{"name":"2014 IEEE Visual Communications and Image Processing Conference","volume":"105 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116420267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信