2017 IEEE International Conference on Multimedia and Expo (ICME)最新文献

筛选
英文 中文
Edge-preserving disparity map estimation from stereo videos for bokeh synthesis 基于散景合成的立体视频保边视差图估计
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019534
Wei-Lun Lan, Shih-Hsuan Yao, S. Lai
{"title":"Edge-preserving disparity map estimation from stereo videos for bokeh synthesis","authors":"Wei-Lun Lan, Shih-Hsuan Yao, S. Lai","doi":"10.1109/ICME.2017.8019534","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019534","url":null,"abstract":"We present a new method of estimating disparity maps from stereo videos for bokeh effect synthesis. In this work, we develop an improved total variation regularization and the robust L1 norm in the data fidelity term (TV-L1) [4] based method to estimate edge-preserving disparity map without stereo rectification. The proposed algorithm improves the TV-L1 approach by incorporating structure edge detection, occlusion area detection, textureless region detection and applying the guided filter to alleviate the inconsistency problem between the disparity map and color image around object boundary. Furthermore, we propose a temporal filter to improve the temporal consistency of the disparity maps computed from the stereo videos. We use saliency map to focus the synthesis result on the objects which attract human attention most. Experimental comparisons on various real videos are shown to demonstrate that the proposed algorithm generates more visually pleasing bokeh video synthesis compared with those by using previous stereo matching methods.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"119 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126640985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Residual convolution network based steganalysis with adaptive content suppression 基于残差卷积网络的自适应内容抑制隐写分析
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019304
Songtao Wu, S. Zhong, Yan Liu
{"title":"Residual convolution network based steganalysis with adaptive content suppression","authors":"Songtao Wu, S. Zhong, Yan Liu","doi":"10.1109/ICME.2017.8019304","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019304","url":null,"abstract":"Image steganalysis is to discriminate innocent images and those suspected images with hidden messages. In this paper, we propose a unified Convolutional Neural Network (CNN) model for this task. In order to reliably detect modern steganographic algorithms, we design the proposed model from two aspects. For the first, different from existing CNN based steganalytic algorithms that use a predefined highpass kernel to suppress image content, we integrate the highpass filtering operation into the proposed network by building a content suppression subnetwork. For the second, we propose a novel sub-network to actively preserve the weak stego signal generated by secret messages based on residual learning, making the successive network capture the difference between cover images and stego images. Extensive experiments demonstrate that the proposed model can detect states-of-the-art steganography with much lower detection error rates than previous methods.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116564716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
CP-operated dash caching via reinforcement learning 通过强化学习的cp操作的破折号缓存
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019496
Zhengyuan Pang, Lifeng Sun, Zhi Wang, Wen Hu, Shiqiang Yang
{"title":"CP-operated dash caching via reinforcement learning","authors":"Zhengyuan Pang, Lifeng Sun, Zhi Wang, Wen Hu, Shiqiang Yang","doi":"10.1109/ICME.2017.8019496","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019496","url":null,"abstract":"In recent years, Dynamic Adaptive Streaming over HTTP (DASH) has gained momentum as an effective solution for delivering videos on the Internet. This trend is further driven by the deployment of existing HTTP cache infrastructures in DASH systems to reduce the traffic load as well as to serve clients better. However, deploying conventional cache servers in DASH systems still suffers from low cache hit ratio and bitrate oscillations, which makes it challenging for content providers (CPs) to balance the user-perceived quality-of-experience (QoE) and the operating cost in cache-enabled DASH systems. To address this challenge, we propose a CP-operated DASH caching framework to provide good user QoE with low cost. In particular, we first formulate the caching decision problem as a stochastic optimization problem over a finite time horizon. The objective of this problem is to maximize a weighted sum of the user QoE and the operating cost, termed as the utility. Then we design a reinforcement learning based online algorithm which can obtain approximately optimal solution of this problem. Through extensive trace-driven experiments, we show that our approach not only achieves 40% average improvement of the overall utility compared to baseline approaches, but also adapts to the server load.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122363542","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Saliency map generation based on saccade target theory 基于扫视目标理论的显著性图生成
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019456
Yue Wu, Zhenzhong Chen
{"title":"Saliency map generation based on saccade target theory","authors":"Yue Wu, Zhenzhong Chen","doi":"10.1109/ICME.2017.8019456","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019456","url":null,"abstract":"In this paper, a novel saliency map generation approach based on the saccade target theory is proposed. A probabilistic model of transsaccadic integration is built based on four cues that influence human visual attention: foveaperiphery resolution discrepancy, visual memory, oculomotor bias and inhibition of return (IOR), where visual memory is formulated as combination of the visual short-term memory (VSTM) and the visual long-term memory (VLTM). A sequence of fixations is generated based on the model to simulate the shift of attention. The proposed approach is evaluated on both perspectives of the accuracy of generated saliency maps and scanpaths. As demonstrated in the experiments, our method provides better estimated saliency maps and scanpaths on eye tracking datasets when compared to several state-of-the-art models.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114270854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Feature guided non-rigid image/surface deformation via moving least squares with manifold regularization 通过流形正则化的移动最小二乘,特征引导非刚性图像/表面变形
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019423
Huabing Zhou, Jiayi Ma, Yanduo Zhang, Zhenghong Yu, Shiqiang Ren, Deng Chen
{"title":"Feature guided non-rigid image/surface deformation via moving least squares with manifold regularization","authors":"Huabing Zhou, Jiayi Ma, Yanduo Zhang, Zhenghong Yu, Shiqiang Ren, Deng Chen","doi":"10.1109/ICME.2017.8019423","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019423","url":null,"abstract":"In this paper, a novel closed-form transformation estimation method based on feature guided moving least squares together with manifold regularization is proposed for nonrigid image/surface deformation. The method takes the user-controlled point-offset-vectors and the feature points of the image/surface as input, and estimates the spatial transformation between the two control point sets for each pixel/voxel. To achieve a detail-preserving and realistic deformation, the transformation estimation is formulated as a vector-field interpolation problem using a feature guided moving least squares method, where a manifold regularization is imposed as a prior on the transformation to capture the underlying intrinsic geometry of the input image/surface. The non-rigid transformation is specified in a reproducing kernel Hilbert space. We derive a closed-form solution of the transformation and adopt a sparse approximation to achieve a fast implementation, which largely reduces the computation complexity without performance sacrifice. In addition, the proposed method can give a wonderful user experience, fast and convenient manipulating. Extensive experiments on both 2D and 3D data demonstrate that the proposed method can produce more natural deformations compared with other state-of-the-art methods.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122172017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Scanpath mining of eye movement trajectories for visual attention analysis 用于视觉注意分析的眼动轨迹扫描路径挖掘
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019507
Aoqi Li, Yingxue Zhang, Zhenzhong Chen
{"title":"Scanpath mining of eye movement trajectories for visual attention analysis","authors":"Aoqi Li, Yingxue Zhang, Zhenzhong Chen","doi":"10.1109/ICME.2017.8019507","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019507","url":null,"abstract":"Eye movement reflects the shift of overt visual attention. Eye movement trajectories from a group of observers can be expressed by a representative scanpath. The representative scan-path can work as a baseline for studies on scanpath prediction as well as provide useful knowledge about group behavior in psychological studies. In this paper, we propose a new framework to summarize a representative scanpath from individual scanpaths, taking into account the spatial distribution of scan-paths rather than simply treating them as strings of characters. It consists of three steps: extract areas of interest (AOI), remove outliers and summarize scanpaths. In the last step, we develop an algorithm termed Candidate-constrained DTW Barycenter Algorithm (CDBA) by imposing 3 constraints: (1) the components of the representative scanpath must be chosen from candidates (extracted AOIs); (2) the occurrence count of each AOI in the representative scanpath cannot exceed its maximum occurrence count in individual scanpaths; (3) any two contiguous AOIs in the representative scanpath must be contiguous in at least one individual scanpath. The experiments demonstrate that the proposed method outperforms other state-of-the-art scanpath mining methods.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130396197","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Steered mixture-of-experts for light field coding, depth estimation, and processing 操纵混合专家光场编码,深度估计和处理
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019442
Ruben Verhack, T. Sikora, Lieven Lange, Rolf Jongebloed, G. Wallendael, P. Lambert
{"title":"Steered mixture-of-experts for light field coding, depth estimation, and processing","authors":"Ruben Verhack, T. Sikora, Lieven Lange, Rolf Jongebloed, G. Wallendael, P. Lambert","doi":"10.1109/ICME.2017.8019442","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019442","url":null,"abstract":"The proposed framework, called Steered Mixture-of-Experts (SMoE), enables a multitude of processing tasks on light fields using a single unified Bayesian model. The underlying assumption is that light field rays are instantiations of a non-linear or non-stationary random process that can be modeled by piecewise stationary processes in the spatial domain. As such, it is modeled as a space-continuous Gaussian Mixture Model. Consequently, the model takes into account different regions of the scene, their edges, and their development along the spatial and disparity dimensions. Applications presented include light field coding, depth estimation, edge detection, segmentation, and view interpolation. The representation is compact, which allows for very efficient compression yielding state-of-the-art coding results for low bit-rates. Furthermore, due to the statistical representation, a vast amount of information can be queried from the model even without having to analyze the pixel values. This allows for “blind” light field processing and classification.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126686660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Estimating relative depth in single images via rankboost 通过rankboost估计单幅图像的相对深度
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019434
R. Ewerth, Matthias Springstein, Eric Müller, Alexander Balz, Jan Gehlhaar, T. Naziyok, K. Dembczynski, E. Hüllermeier
{"title":"Estimating relative depth in single images via rankboost","authors":"R. Ewerth, Matthias Springstein, Eric Müller, Alexander Balz, Jan Gehlhaar, T. Naziyok, K. Dembczynski, E. Hüllermeier","doi":"10.1109/ICME.2017.8019434","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019434","url":null,"abstract":"In this paper, we present a novel approach to estimate the relative depth of regions in monocular images. There are several contributions. First, the task of monocular depth estimation is considered as a learning-to-rank problem which offers several advantages compared to regression approaches. Second, monocular depth clues of human perception are modeled in a systematic manner. Third, we show that these depth clues can be modeled and integrated appropriately in a Rankboost framework. For this purpose, a space-efficient version of Rankboost is derived that makes it applicable to rank a large number of objects, as posed by the given problem. Finally, the monocular depth clues are combined with results from a deep learning approach. Experimental results show that the error rate is reduced by adding the monocular features while outperforming state-of-the-art systems.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"140 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123386524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Pixel-level guided face editing with fully convolution networks 像素级引导的面部编辑与完全卷积网络
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019363
Zhenxi Li, Juyong Zhang
{"title":"Pixel-level guided face editing with fully convolution networks","authors":"Zhenxi Li, Juyong Zhang","doi":"10.1109/ICME.2017.8019363","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019363","url":null,"abstract":"Face editing has a variety of applications, especially with the increasing popularity of photography using mobile devices. In this work, we argue that the performance of face image editing can be further improved by using semantic segmentation which marks each pixel with a label that indicates its corresponding facial part. To this end, we propose a deep learning based method for automatic pixel-level labeling on face images. Our approach achieves state-of-the-art labeling accuracy on publicly available datasets, at a significantly higher speed than existing labeling methods. Then we show how the label information can be applied to various face image editing applications, such as face smoothing, face cloning and face blending. Extensive experimental results demonstrate the effectiveness of our method in editing face images with convincing visual quality.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124439772","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Hyper-spectral image reconstruction based on SL0-SL0 minimization 基于SL0-SL0最小化的高光谱图像重建
2017 IEEE International Conference on Multimedia and Expo (ICME) Pub Date : 2017-07-01 DOI: 10.1109/ICME.2017.8019380
Xinyue Zhang, Xudong Zhang
{"title":"Hyper-spectral image reconstruction based on SL0-SL0 minimization","authors":"Xinyue Zhang, Xudong Zhang","doi":"10.1109/ICME.2017.8019380","DOIUrl":"https://doi.org/10.1109/ICME.2017.8019380","url":null,"abstract":"This paper proposes a new prior image constrained compressive sampling (PICCS) method to reconstruct hyper-spectral images, namely SL0-SL0 minimization-based hyper-spectral imaging (HSI). This is a band-by-band reconstruction method, which reconstructs each hyper-spectral band based on the previous one. This method utilizes not only the sparsity of each hyper-spectral band in certain bases but also the similarity between two consecutive bands. In addition, compared with the popular approaches which reconstruct all the hyper-spectral bands simultaneously, SL0-SL0 minimization-based HSI reduce the requirements to computational ability and memory of receivers for that only one hyper-spectral band is reconstructed at each time. Compared with the exiting PICCS methods, which lose efficiency to reconstruct signals with large size, the SL0-SL0 minimization method significantly speeds up the reconstruction procedure. Some simulations are provided to illustrate the effectiveness of the proposed method.","PeriodicalId":330977,"journal":{"name":"2017 IEEE International Conference on Multimedia and Expo (ICME)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127706948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信