2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)最新文献

筛选
英文 中文
Periodic Variance Maximization Using Generalized Eigenvalue Decomposition Applied to Remote Photoplethysmography Estimation 基于广义特征值分解的周期方差最大化方法在光体积脉搏波遥感估计中的应用
R. Macwan, Serge Bobbia, Y. Benezeth, Julien Dubois, A. Mansouri
{"title":"Periodic Variance Maximization Using Generalized Eigenvalue Decomposition Applied to Remote Photoplethysmography Estimation","authors":"R. Macwan, Serge Bobbia, Y. Benezeth, Julien Dubois, A. Mansouri","doi":"10.1109/CVPRW.2018.00181","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00181","url":null,"abstract":"A generic periodic variance maximization algorithm to extract periodic or quasi-periodic signals of unknown periods embedded into multi-channel temporal signal recordings is described in this paper. The algorithm combines the notion of maximizing a periodicity metric combined with the global optimization scheme to estimate the source periodic signal of an unknown period. The periodicity maximization is performed using Generalized Eigenvalue Decomposition (GEVD) and the global optimization is performed using tabu search. A case study of remote photoplethysmography signal estimation has been utilized to assess the performance of the method using videos from public databases UBFC-RPPG [1] and MMSE-HR [31]. The results confirm the improved performance over existing state of the art methods and the feasibility of the use of the method in a live scenario owing to its small execution time.","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129351798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Camera Calibration for Underwater 3D Reconstruction Based on Ray Tracing Using Snell’s Law 基于Snell定律光线追踪的水下三维重建摄像机标定
Malte Pedersen, S. Bengtson, Rikke Gade, N. Madsen, T. Moeslund
{"title":"Camera Calibration for Underwater 3D Reconstruction Based on Ray Tracing Using Snell’s Law","authors":"Malte Pedersen, S. Bengtson, Rikke Gade, N. Madsen, T. Moeslund","doi":"10.1109/CVPRW.2018.00190","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00190","url":null,"abstract":"Accurately estimating the 3D position of underwater objects is of great interest when doing research on marine animals. An inherent problem of 3D reconstruction of underwater positions is the presence of refraction which invalidates the assumption of a single viewpoint. Three ways of performing 3D reconstruction on underwater objects are compared in this work: an approach relying solely on in-air camera calibration, an approach with the camera calibration performed under water and an approach based on ray tracing with Snell's law. As expected, the in-air camera calibration showed to be the most inaccurate as it does not take refraction into account. The precision of the estimated 3D positions based on the underwater camera calibration and the ray tracing based approach were, on the other hand, almost identical. However, the ray tracing based approach is found to be advantageous as it is far more flexible in terms of the calibration procedure due to the decoupling of the intrinsic and extrinsic camera parameters.","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127455546","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Multi-scale Single Image Dehazing Using Perceptual Pyramid Deep Network 基于感知金字塔深度网络的多尺度单幅图像去雾
He Zhang, Vishwanath A. Sindagi, Vishal M. Patel
{"title":"Multi-scale Single Image Dehazing Using Perceptual Pyramid Deep Network","authors":"He Zhang, Vishwanath A. Sindagi, Vishal M. Patel","doi":"10.1109/CVPRW.2018.00135","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00135","url":null,"abstract":"Haze adversely degrades quality of an image thereby affecting its aesthetic appeal and visibility in outdoor scenes. Single image dehazing is particularly challenging due to its ill-posed nature. Most existing work, including the recent convolutional neural network (CNN) based methods, rely on the classical mathematical formulation where the hazy image is modeled as the superposition of attenuated scene radiance and the atmospheric light. In this work, we explore CNNs to directly learn a non-linear function between hazy images and the corresponding clear images. We present a multi-scale image dehazing method using Perceptual Pyramid Deep Network based on the recently popular dense blocks and residual blocks. The proposed method involves an encoder-decoder structure with a pyramid pooling module in the decoder to incorporate contextual information of the scene while decoding. The network is learned by minimizing the mean squared error and perceptual losses. Multi-scale patches are used during training and inference process to further improve the performance. Experiments on the recently released NTIRE2018-Dehazing dataset demonstrates the superior performance of the proposed method over recent state-of-the-art approaches. Additionally, the proposed method is ranked among top-3 methods in terms of quantitative performance in the recently conducted NTIRE2018-Dehazing challenge. Code can be found at https://github.com/hezhangsprinter/NTIRE-2018-Dehazing-Challenge","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132212077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 128
Elastic Handling of Predictor Phase in Functional Regression Models 函数回归模型中预测期的弹性处理
Kyungmin Ahn, J. D. Tucker, Wei Wu, Anuj Srivastava
{"title":"Elastic Handling of Predictor Phase in Functional Regression Models","authors":"Kyungmin Ahn, J. D. Tucker, Wei Wu, Anuj Srivastava","doi":"10.1109/CVPRW.2018.00072","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00072","url":null,"abstract":"Functional variables serve important roles as predictors in a variety of pattern recognition and vision applications. Focusing on a specific subproblem, termed scalar-on-function regression, most current approaches adopt the standard L2 inner product to form a link between functional predictors and scalar responses. These methods may perform poorly when predictor functions contain nuisance phase variability, i.e., predictors are temporally misaligned due to noise. While a simple solution could be to prealign predictors as a pre-processing step, before applying a regression model, this alignment is seldom optimal from the perspective of regression. We propose a new approach, termed elastic functional regression, where alignment is included in the regression model itself, and is performed in conjunction with the estimation of other model parameters. This model is based on a norm-preserving warping of predictors, not the standard time warping of functions, and provides better prediction in situations where the shape or the amplitude of the predictor is more useful than its phase. We demonstrate the effectiveness of this framework using simulated and stock market data.","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129222184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Active Vision Dataset Benchmark 主动视觉数据集基准
Phil Ammirato, A. Berg, J. Kosecka
{"title":"Active Vision Dataset Benchmark","authors":"Phil Ammirato, A. Berg, J. Kosecka","doi":"10.1109/CVPRW.2018.00277","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00277","url":null,"abstract":"Several recent efforts in computer vision indicate a trend toward studying and understanding problems in larger scale environments, beyond single images, and focus on connections to tasks in navigation, mobile manipulation, and visual question answering. A common goal of these tasks is the capability of moving in the environment, acquiring novel views during perception and while performing a task. This capability comes easily in synthetic environments, however achieving the same effect with real images is much more laborious. We propose using the existing Active Vision Dataset to form a benchmark for such problems in a real-world settings with real images. The dataset is well suited for evaluating tasks of multiview active recognition, target driven navigation, and target search, and also can be effective for studying the transfer of strategies learned in simulation to real settings.","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125480528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Novel Framework for Remote Photoplethysmography Pulse Extraction on Compressed Videos 一种基于压缩视频的光容积脉搏波脉冲提取新框架
Changchen Zhao, Chun-Liang Lin, Weihai Chen, Zhengguo Li
{"title":"A Novel Framework for Remote Photoplethysmography Pulse Extraction on Compressed Videos","authors":"Changchen Zhao, Chun-Liang Lin, Weihai Chen, Zhengguo Li","doi":"10.1109/CVPRW.2018.00177","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00177","url":null,"abstract":"Remote photoplethysmography (rPPG) has recently attracted much attention due to its non-contact measurement convenience and great potential in health care and computer vision applications. However, almost all the existing rPPG methods are based on uncompressed video data, which greatly limits its application to the scenarios that require long-distance video transmission. This paper proposes a novel framework as a first attempt to address the rPPG pulse extraction in presence of video compression artifacts. Based on the analysis of the impact of various compression methods on rPPG measurements, the problem is cast as single-channel signal separation. The framework consists of three major steps to extract the pulse waveform and heart rate by exploiting frequency structure of the rPPG signal. A benchmark dataset which contains stationary and motion videos has been built. The results show that the proposed algorithm significantly improves the SNR and heart rate precision of state-of-the-art rPPG algorithms on stationary videos and has a positive effect on motion videos at low bitrates.","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123368753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Cross-Domain Fashion Image Retrieval 跨域时尚图像检索
B. Gajic, R. Baldrich
{"title":"Cross-Domain Fashion Image Retrieval","authors":"B. Gajic, R. Baldrich","doi":"10.1109/CVPRW.2018.00243","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00243","url":null,"abstract":"Cross domain image retrieval is a challenging task that implies matching images from one domain to their pairs from another domain. In this paper we focus on fashion image retrieval, which involves matching an image of a fashion item taken by users, to the images of the same item taken in controlled condition, usually by professional photographer. When facing this problem, we have different products in train and test time, and we use triplet loss to train the network. We stress the importance of proper training of simple architecture, as well as adapting general models to the specific task.","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121208438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
On Visible Adversarial Perturbations & Digital Watermarking 可见对抗性扰动与数字水印
Jamie Hayes
{"title":"On Visible Adversarial Perturbations & Digital Watermarking","authors":"Jamie Hayes","doi":"10.1109/CVPRW.2018.00210","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00210","url":null,"abstract":"Given a machine learning model, adversarial perturbations transform images such that the model's output is classified as an attacker chosen class. Most research in this area has focused on adversarial perturbations that are imperceptible to the human eye. However, recent work has considered attacks that are perceptible but localized to a small region of the image. Under this threat model, we discuss both defenses that remove such adversarial perturbations, and attacks that can bypass these defenses.","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"40 15","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113936068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 92
Generative Adversarial Style Transfer Networks for Face Aging 面部老化的生成对抗风格迁移网络
Sveinn Pálsson, E. Agustsson, R. Timofte, L. Gool
{"title":"Generative Adversarial Style Transfer Networks for Face Aging","authors":"Sveinn Pálsson, E. Agustsson, R. Timofte, L. Gool","doi":"10.1109/CVPRW.2018.00282","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00282","url":null,"abstract":"How somebody looked like when younger? What could a person look like when 10 years older? In this paper we look at the problem of face aging, which relates to processing an image of a face to change its apparent age. This task involves synthesizing images and modeling the aging process, which both are problems that have recently enjoyed much research interest in the field of face and gesture recognition. We propose to look at the problem from the perspective of image style transfer, where we consider the age of the person as the underlying style of the image. We show that for large age differences, convincing face aging can be achieved by formulating the problem with a pairwise training of Cycle-consistent Generative Adversarial Networks (CycleGAN) over age groups. Furthermore, we propose a variant of CycleGAN which directly incorporates a pre-trained age prediction model, which performs better when the desired age difference is smaller. The proposed approaches are complementary in strengths and their fusion performs well for any desired level of aging effect. We quantitatively evaluate our proposed method through a user study and show that it outperforms prior state-of-the-art techniques for face aging.","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124912787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 47
A Directed Sparse Graphical Model for Multi-target Tracking 多目标跟踪的有向稀疏图模型
M. Ullah, F. A. Cheikh
{"title":"A Directed Sparse Graphical Model for Multi-target Tracking","authors":"M. Ullah, F. A. Cheikh","doi":"10.1109/CVPRW.2018.00235","DOIUrl":"https://doi.org/10.1109/CVPRW.2018.00235","url":null,"abstract":"We propose a Directed Sparse Graphical Model (DSGM) for multi-target tracking. In the category of global optimization for multi-target tracking, traditional approaches have two main drawbacks. First, a cost function is defined in terms of the linear combination of the spatial and appearance constraints of the targets which results a highly non-convex function. And second, a very dense graph is constructed to capture the global attribute of the targets. In such a graph, It is impossible to find reliable tracks in polynomial time unless some relaxation and heuristics are used. To address these limitations, we proposed DSGM which finds a set of reliable tracks for the targets without any heuristics or relaxation and keeps the computational complexity very low through the design of the graph. Irrespective of traditional approaches where spatial and appearance constraints are added up linearly with a given weight factor, we incorporated these constraints in a cascaded fashion. First, we exploited a Hidden Markov Model (HMM) for the spatial constraints of the target and obtain most probable locations of the targets in a segment of video. Afterwards, a deep feature based appearance model is used to generate the sparse graph. The track for each target is found through dynamic programming. Experiments are performed on 3 challenging sports datasets (football, basketball and sprint) and promising results are achieved.","PeriodicalId":150600,"journal":{"name":"2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"39 25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114053380","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 54
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信