2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)最新文献

筛选
英文 中文
Performance Analysis of WebRTC Embedding Optimized HEVC CodeC 嵌入优化HEVC编解码器的WebRTC性能分析
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008812
Zhenyu Liu
{"title":"Performance Analysis of WebRTC Embedding Optimized HEVC CodeC","authors":"Zhenyu Liu","doi":"10.1109/VCIP56404.2022.10008812","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008812","url":null,"abstract":"Web Real-Time Communications (WebRTC) is an open-source platform, supporting developer to build voice- and video-communication solutions. Video compression engine is one infrastructure in WebRTC. As video stream accounting for more than 90% bandwidth requirement in real-time communication. The performance of embedded video encoder in WebRTC determines the Quality of Experience of communication, including the picture quality, the communication latency and the playback smoothness. In our WebRTC, we implemented the high performance HEVC software encoder as a substitute of the VP8 encoder in WebRTC. We improve the compression efficiency of HEVC encoder significantly. As compared with the most advanced x265, the averaged 34.6% rate saving was achieved by our encoder, especially in low-latency coding applications. As compared to the default WebRTC integrated with VP8, up to 90.2% bandwidth requirement was saved by our optimizations.","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126724114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Intelligent Reflection Elimination Imaging Device based on Polarizer 基于偏光镜的智能反射消除成像装置
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008900
Xinru Chen, Hang Liu, Menghan Hu, Lejing Zhang, Yunmian Li
{"title":"Intelligent Reflection Elimination Imaging Device based on Polarizer","authors":"Xinru Chen, Hang Liu, Menghan Hu, Lejing Zhang, Yunmian Li","doi":"10.1109/VCIP56404.2022.10008900","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008900","url":null,"abstract":"Glass reflection is a problem when taking photos through glass windows or showcases. As the visual quality of captured image can be enhanced by removing reflection, we develop an intelligent reflection elimination imaging device based on polarizer to minimize reflection effect on the images. The system mainly consists of a polarizing module, an image analysis module and a reflection removal module. The users can hold the device and capture images with minimum reflection whether in the day or night. The demo video is available at: https://doi.org/10.6084/m9.figshare.19687830.v1.","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"255 5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116574047","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High-frequency guided CNN for video compression artifacts reduction 高频导引CNN用于视频压缩伪影的减少
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008814
Li Yu, Wenshuai Chang, Qingshan Liu, M. Gabbouj
{"title":"High-frequency guided CNN for video compression artifacts reduction","authors":"Li Yu, Wenshuai Chang, Qingshan Liu, M. Gabbouj","doi":"10.1109/VCIP56404.2022.10008814","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008814","url":null,"abstract":"In this paper, we propose a high-frequency guided CNN for video compression artifacts reduction. In the proposed method, high frequency component in Y channel is extracted and used to guide the quality enhancement of all Y, U, V channels. As high frequency component contains the edge and contour information of the objects in the image, which is of vital importance to both subjective and objective quality. In general, the proposed method consists of two modules: the high frequency guidance module and the quality enhancement module. The high-frequency guidance module uses multiple octave convolutions to extract the high-frequency component in Y channel and then fuse it into the features of Y, U, and V channels. While in the quality enhancement module, multiple CNN residual blocks are used for the quality enhancement of Y, U, and V channels. The proposed method was integrated into both HM-16.22 and VTM-16.0. The results on the JVET test sequence under All Intra configuration shows the effectiveness of the proposed method. Compared with HEVC, the proposed method achieves the average BD-rate reductions of -12.3%, -22.7% and -23.5% for Y, U and V channels respectively. Compared with VVC, the average BD-rate reductions are -6.7%, -12.3% and -13.2% correspondingly.","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116624078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Controllable Space-Time Video Super-Resolution via Enhanced Bidirectional Flow Warping 基于增强双向流扭曲的可控时空视频超分辨率
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008838
Yuantong Zhang, Huairui Wang, Zhenzhong Chen
{"title":"Controllable Space-Time Video Super-Resolution via Enhanced Bidirectional Flow Warping","authors":"Yuantong Zhang, Huairui Wang, Zhenzhong Chen","doi":"10.1109/VCIP56404.2022.10008838","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008838","url":null,"abstract":"Space-time video super-resolution targets to increase a given video's frame rate and resolution simultaneously. Al-though existing approaches have made great progress, most of them still suffer from the inaccurate approximation of large motions or fail to generate temporal consistent motion trajectory. To alleviate these problems, we carefully review the characteris-tics of different optical flow warping strategies, integrating and enhancing them to achieve more robust capabilities for handling extreme motions and time-modulated interpolation. Specifically, we utilize enhanced backward warping to perform alignment, mine space-time information across low resolution input frames, and propose an enhanced forward warping strategy to interpolate arbitrary intermediate frames. Furthermore, the proposed model can be trained end-to-end and produce intermediate results at any time by merely supervising the center moment. Experimental results show that the proposed algorithm performs favorably against the state-of-the-art methods in objective metrics and subjective visual effects.","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131104239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Robust Dynamic Background Modeling for Foreground Estimation 前景估计的鲁棒动态背景建模
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008816
Qiang Ning, Fangfang Wu, W. Dong, Jinjian Wu, Guangming Shi, Xin Li
{"title":"Robust Dynamic Background Modeling for Foreground Estimation","authors":"Qiang Ning, Fangfang Wu, W. Dong, Jinjian Wu, Guangming Shi, Xin Li","doi":"10.1109/VCIP56404.2022.10008816","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008816","url":null,"abstract":"Separating the background and foreground components from video frames is important to many tasks in computer vision and multimedia. As of today, robust principal component analysis (RPCA) has shown highly promising performance with the assumption that the background is low-rank and the foreground is sparse. However, existing RPCA-based methods have overlooked the uncertainty that some parts of the background (e.g., moving leaves in a dynamic background) or even the whole background (e.g., camera jittering) can be moving, which violates the low-rank assumption. To address this issue, we propose a novel enhanced RPCA framework (called ERPCA) by robustly modeling the dynamic background. Different from traditional RPCA framework, the background is decomposed into a low-rank component and a sparse component in the proposed ERPCA framework. Specifically, the sparse parts including foreground and dynamic parts of the background are modeled by Gaussian scale mixture (GSM) model. Moreover, those sparse components are further constrained by temporal consistency using nonzeromeans Gaussian models; the correspondences between sparse pixels in adjacent frames are explored by optical flow. Experimental results on 40 real videos demonstrate the superiority of our proposed method, with better average results than current state-of-the-art foreground estimation methods.","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124101580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Asynchronous Autoregressive Prediction for Satellite Anomaly Detection 卫星异常检测的异步自回归预测
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008889
Peng Liu, Haopeng Zhang, Lifang Yuan, Borui Zhang, Chengkun Wang
{"title":"Asynchronous Autoregressive Prediction for Satellite Anomaly Detection","authors":"Peng Liu, Haopeng Zhang, Lifang Yuan, Borui Zhang, Chengkun Wang","doi":"10.1109/VCIP56404.2022.10008889","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008889","url":null,"abstract":"This paper proposes an ASynchronous Autoregressive Prediction (ASAP) method for satellite anomaly detection. We empirically observe that a single classification model can hardly detect unknown anomalous situations and neglect the Markov nature of temporal satellite data. To address this, we adopt an autoregressive model to deal with the prediction of unknown anomaly for satellite data. We further propose a non-uniform temporal encoding method for asynchronous data and a median filtering method for more accurate detection. To reduce the effect of outliers, we employ an adaptive threshold selection method to achieve a more robust classification boundary. Experiments on real satellite data demonstrate that the proposed ASAP method outperforms the baseline classification method by 55.79%.","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"508 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117037975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Augmented Normalizing Flow for Point Cloud Geometry Coding 点云几何编码的增强归一化流程
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008821
Siao-Yu Li, Ji-Jin Chiu, J. Chiang, Wen-Hsiao Peng, W. Lie
{"title":"Augmented Normalizing Flow for Point Cloud Geometry Coding","authors":"Siao-Yu Li, Ji-Jin Chiu, J. Chiang, Wen-Hsiao Peng, W. Lie","doi":"10.1109/VCIP56404.2022.10008821","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008821","url":null,"abstract":"With the increased popularity of immersive media, point clouds have become one of the popular data representations for presenting 3D scenes. The huge amount of point cloud data poses a great challenge on their storage and real-time transmission, which calls for efficient point cloud compression. This paper presents a novel point cloud geometry compression technique based on learning end-to-end an augmented normalizing flow (ANF) model to represent the occupancy status of voxelized data points. The higher expressive power of ANF than variational autoencoders (V AE) is leveraged for the first time to represent binary occupancy status. Compared to two coding standards developed by MPEG, namely G-PCC (geometry-based point cloud compression) and V-PCC (video-based point cloud compression), our method achieves more than 80% and 30% bitrate reduction, respectively. Compared to several learning-based methods, our method also yields better performance.","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"80 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133901727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
ENDE-GNN: An Encoder-decoder GNN Framework for Sketch Semantic Segmentation ENDE-GNN:一个用于草图语义分割的编码器-解码器GNN框架
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008880
Yixiao Zheng, Jiyang Xie, Aneeshan Sain, Zhanyu Ma, Yi-Zhe Song, Jun Guo
{"title":"ENDE-GNN: An Encoder-decoder GNN Framework for Sketch Semantic Segmentation","authors":"Yixiao Zheng, Jiyang Xie, Aneeshan Sain, Zhanyu Ma, Yi-Zhe Song, Jun Guo","doi":"10.1109/VCIP56404.2022.10008880","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008880","url":null,"abstract":"Sketch semantic segmentation serves as an important part of sketch interpretation. Recently, some researchers have obtained significant results using graph neural networks (GNN) for this task. However, existing GNN-based methods usually neglect the drawing order of sketches thus missing out the sequence information inherent to sketches. Towards solving this problem to achieve better performance on sketch semantic segmentation, we propose an encoder-decoder GNN framework named ENDE-GNN. Working with an auxiliary decoder, our ENDE-GNN guides the GNN backbone network to not only extract the inter-stroke and intra-stroke features, but also pays attention to the drawing order of sketches. This decoder acts during training only, preventing any additional overhead during testing. The proposed ENDE-GNN obtains state-of-the-art per-formances on three public sketch semantic segmentation datasets, namely SPG, SketchSeg-150K, and CreativeSketch. We further evaluate the effectiveness of ENDE-GNN via ablation studies and visualizations. Codes are available at https://github.com/PRIS-CV/ENDE_For_SSS.","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"521 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134189944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Portable Eye Movement Feature Collection Device for Children with Autism 自闭症儿童便携式眼动特征采集装置
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008848
Xinding Xia, Menghan Hu, Xiaojuan Xue, Qiaoyun Liu, Jian Zhang, Guangtao Zhai
{"title":"Portable Eye Movement Feature Collection Device for Children with Autism","authors":"Xinding Xia, Menghan Hu, Xiaojuan Xue, Qiaoyun Liu, Jian Zhang, Guangtao Zhai","doi":"10.1109/VCIP56404.2022.10008848","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008848","url":null,"abstract":"Eye movement data has become an important char-acterization in the analysis of children with autism spectrum disorder (ASD). Current eye movement measurement meth-ods require specialized expensive equipment, calibration, and trained personnel, limiting their use in general ASD screening, especially in resource-scarce environments. Therefore, collecting eye movement features based on the standard RGB camera of a mobile phone or tablet has many advantages over professional equipment. The system design is based on the Android tablet design, and the screen is divided into two parts to display the normal children and the ASD children paintings. The eye movement data of children is obtained through the front camera, so as to provide data support for future data analysis. Taking the different cooperation degrees of children into account, two collection modes are designed: 1) directly displaying the stimuli in a loop (image mode); and 2) providing the background video interspersed with the stimulus display (video mode). The demo video of the proposed system is available at: https://doi.org/10.6084/m9.figshare.21346806.v1.","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"21 12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131681616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MCascade R-CNN: A Modified Cascade R-CNN for Detection of Calcified on Coronary Artery Angiography Images MCascade R-CNN:一种改进的级联R-CNN检测冠状动脉造影图像钙化
2022 IEEE International Conference on Visual Communications and Image Processing (VCIP) Pub Date : 2022-12-13 DOI: 10.1109/VCIP56404.2022.10008804
W. Wang, Yi Zhang, Xiaofei Wang, Honggang Zhang, Lihua Xie, Bo Xu
{"title":"MCascade R-CNN: A Modified Cascade R-CNN for Detection of Calcified on Coronary Artery Angiography Images","authors":"W. Wang, Yi Zhang, Xiaofei Wang, Honggang Zhang, Lihua Xie, Bo Xu","doi":"10.1109/VCIP56404.2022.10008804","DOIUrl":"https://doi.org/10.1109/VCIP56404.2022.10008804","url":null,"abstract":"Among cardiovascular diseases, coronary artery calcification (CAC) is a high-risk factor for worsening protopathy and increased mortality. However, the coronary artery an-giogram, which is the main approach for CAC diagnosis, suffers from plenty of photographing noise. This brings difficulties to detect calcification from the background. In this paper, a modified Cascade R-CNN (MCascade R-CNN) network is proposed to deal with the problem of calcium detection in angiograms. In the proposed network, we propose an innovative balanced aggregation pyramid structure, integrating multi-level features of every depth in the feature map, based on enhanced propagation of strong semantic features. In addition, a new convolutional attention mechanism is designed to improve the performance of the detector. Experiments show that the proposed method enjoys better performance in detecting and marking CAC in angiograms,","PeriodicalId":269379,"journal":{"name":"2022 IEEE International Conference on Visual Communications and Image Processing (VCIP)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123022148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信