2012 IEEE Conference on Computer Vision and Pattern Recognition最新文献

筛选
英文 中文
Unsupervised feature learning framework for no-reference image quality assessment 无参考图像质量评估的无监督特征学习框架
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6247789
Peng Ye, J. Kumar, Le Kang, D. Doermann
{"title":"Unsupervised feature learning framework for no-reference image quality assessment","authors":"Peng Ye, J. Kumar, Le Kang, D. Doermann","doi":"10.1109/CVPR.2012.6247789","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6247789","url":null,"abstract":"In this paper, we present an efficient general-purpose objective no-reference (NR) image quality assessment (IQA) framework based on unsupervised feature learning. The goal is to build a computational model to automatically predict human perceived image quality without a reference image and without knowing the distortion present in the image. Previous approaches for this problem typically rely on hand-crafted features which are carefully designed based on prior knowledge. In contrast, we use raw-image-patches extracted from a set of unlabeled images to learn a dictionary in an unsupervised manner. We use soft-assignment coding with max pooling to obtain effective image representations for quality estimation. The proposed algorithm is very computationally appealing, using raw image patches as local descriptors and using soft-assignment for encoding. Furthermore, unlike previous methods, our unsupervised feature learning strategy enables our method to adapt to different domains. CORNIA (Codebook Representation for No-Reference Image Assessment) is tested on LIVE database and shown to perform statistically better than the full-reference quality measure, structural similarity index (SSIM) and is shown to be comparable to state-of-the-art general purpose NR-IQA algorithms.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115117935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 710
Hierarchical matching with side information for image classification 基于侧信息的图像分类层次匹配
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6248083
Qiang Chen, Zheng Song, Yang Hua, Zhongyang Huang, Shuicheng Yan
{"title":"Hierarchical matching with side information for image classification","authors":"Qiang Chen, Zheng Song, Yang Hua, Zhongyang Huang, Shuicheng Yan","doi":"10.1109/CVPR.2012.6248083","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6248083","url":null,"abstract":"In this work, we introduce a hierarchical matching framework with so-called side information for image classification based on bag-of-words representation. Each image is expressed as a bag of orderless pairs, each of which includes a local feature vector encoded over a visual dictionary, and its corresponding side information from priors or contexts. The side information is used for hierarchical clustering of the encoded local features. Then a hierarchical matching kernel is derived as the weighted sum of the similarities over the encoded features pooled within clusters at different levels. Finally the new kernel is integrated with popular machine learning algorithms for classification purpose. This framework is quite general and flexible, other practical and powerful algorithms can be easily designed by using this framework as a template and utilizing particular side information for hierarchical clustering of the encoded local features. To tackle the latent spatial mismatch issues in SPM, we design in this work two exemplar algorithms based on two types of side information: object confidence map and visual saliency map, from object detection priors and within-image contexts respectively. The extensive experiments over the Caltech-UCSD Birds 200, Oxford Flowers 17 and 102, PASCAL VOC 2007, and PASCAL VOC 2010 databases show the state-of-the-art performances from these two exemplar algorithms.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116900696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 97
Discriminative feature fusion for image classification 判别特征融合用于图像分类
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6248084
Basura Fernando, É. Fromont, Damien Muselet, M. Sebban
{"title":"Discriminative feature fusion for image classification","authors":"Basura Fernando, É. Fromont, Damien Muselet, M. Sebban","doi":"10.1109/CVPR.2012.6248084","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6248084","url":null,"abstract":"Bag-of-words-based image classification approaches mostly rely on low level local shape features. However, it has been shown that combining multiple cues such as color, texture, or shape is a challenging and promising task which can improve the classification accuracy. Most of the state-of-the-art feature fusion methods usually aim to weight the cues without considering their statistical dependence in the application at hand. In this paper, we present a new logistic regression-based fusion method, called LRFF, which takes advantage of the different cues without being tied to any of them. We also design a new marginalized kernel by making use of the output of the regression model. We show that such kernels, surprisingly ignored so far by the computer vision community, are particularly well suited to achieve image classification tasks. We compare our approach with existing methods that combine color and shape on three datasets. The proposed learning-based feature fusion process clearly outperforms the state-of-the art fusion methods for image classification.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":" 60","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120943203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 90
Image matching using local symmetry features 利用局部对称特征进行图像匹配
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6247677
D. C. Hauagge, Noah Snavely
{"title":"Image matching using local symmetry features","authors":"D. C. Hauagge, Noah Snavely","doi":"10.1109/CVPR.2012.6247677","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6247677","url":null,"abstract":"We present a new technique for extracting local features from images of architectural scenes, based on detecting and representing local symmetries. These new features are motivated by the fact that local symmetries, at different scales, are a fundamental characteristic of many urban images, and are potentially more invariant to large appearance changes than lower-level features such as SIFT. Hence, we apply these features to the problem of matching challenging pairs of photos of urban scenes. Our features are based on simple measures of local bilateral and rotational symmetries computed using local image operations. These measures are used both for feature detection and for computing descriptors. We demonstrate our method on a challenging new dataset containing image pairs exhibiting a range of dramatic variations in lighting, age, and rendering style, and show that our features can improve matching performance for this difficult task.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120961900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 164
Intrinsic shape context descriptors for deformable shapes 可变形形状的内在形状上下文描述符
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6247671
Iasonas Kokkinos, M. Bronstein, R. Litman, A. Bronstein
{"title":"Intrinsic shape context descriptors for deformable shapes","authors":"Iasonas Kokkinos, M. Bronstein, R. Litman, A. Bronstein","doi":"10.1109/CVPR.2012.6247671","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6247671","url":null,"abstract":"In this work, we present intrinsic shape context (ISC) descriptors for 3D shapes. We generalize to surfaces the polar sampling of the image domain used in shape contexts: for this purpose, we chart the surface by shooting geodesic outwards from the point being analyzed; `angle' is treated as tantamount to geodesic shooting direction, and radius as geodesic distance. To deal with orientation ambiguity, we exploit properties of the Fourier transform. Our charting method is intrinsic, i.e., invariant to isometric shape transformations. The resulting descriptor is a meta-descriptor that can be applied to any photometric or geometric property field defined on the shape, in particular, we can leverage recent developments in intrinsic shape analysis and construct ISC based on state-of-the-art dense shape descriptors such as heat kernel signatures. Our experiments demonstrate a notable improvement in shape matching on standard benchmarks.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127183443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 169
A learning-based framework for depth ordering 基于学习的深度排序框架
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6247688
Zhaoyin Jia, Andrew C. Gallagher, Yao-Jen Chang, Tsuhan Chen
{"title":"A learning-based framework for depth ordering","authors":"Zhaoyin Jia, Andrew C. Gallagher, Yao-Jen Chang, Tsuhan Chen","doi":"10.1109/CVPR.2012.6247688","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6247688","url":null,"abstract":"Depth ordering is instrumental for understanding the 3D geometry of an image. Humans are surprisingly good at depth ordering even with abstract 2D line drawings. In this paper we propose a learning-based framework for depth ordering inference. Boundary and junction characteristics are important clues for this task, and we have developed new features based on these attributes. Although each feature individually can produce reasonable depth ordering results, each still has limitations, and we can achieve better performance by combining them. In practice, local depth ordering inferences can be contradictory. Therefore, we propose a Markov Random Field model with terms that are more global than previous work, and use graph optimization to encourage a globally consistent ordering. In addition, to produce better object segmentation for the task of depth ordering, we propose to explicitly enforce closed loops and long edges for the occlusion boundary detection. We collect a new depth-order dataset for this problem, including more than a thousand human-labeled images with various daily objects and configurations. The proposed algorithm shows promising performance over conventional methods on both synthetic and real scenes.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124860194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
A constrained latent variable model 约束潜变量模型
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6247934
Aydin Varol, M. Salzmann, P. Fua, R. Urtasun
{"title":"A constrained latent variable model","authors":"Aydin Varol, M. Salzmann, P. Fua, R. Urtasun","doi":"10.1109/CVPR.2012.6247934","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6247934","url":null,"abstract":"Latent variable models provide valuable compact representations for learning and inference in many computer vision tasks. However, most existing models cannot directly encode prior knowledge about the specific problem at hand. In this paper, we introduce a constrained latent variable model whose generated output inherently accounts for such knowledge. To this end, we propose an approach that explicitly imposes equality and inequality constraints on the model's output during learning, thus avoiding the computational burden of having to account for these constraints at inference. Our learning mechanism can exploit non-linear kernels, while only involving sequential closed-form updates of the model parameters. We demonstrate the effectiveness of our constrained latent variable model on the problem of non-rigid 3D reconstruction from monocular images, and show that it yields qualitative and quantitative improvements over several baselines.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124868527","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 80
Actionable saliency detection: Independent motion detection without independent motion estimation 可操作的显著性检测:独立的运动检测,不需要独立的运动估计
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6247732
Georgios Georgiadis, Alper Ayvaci, Stefano Soatto
{"title":"Actionable saliency detection: Independent motion detection without independent motion estimation","authors":"Georgios Georgiadis, Alper Ayvaci, Stefano Soatto","doi":"10.1109/CVPR.2012.6247732","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6247732","url":null,"abstract":"We present a model and an algorithm to detect salient regions in video taken from a moving camera. In particular, we are interested in capturing small objects that move independently in the scene, such as vehicles and people as seen from aerial or ground vehicles. Many of the scenarios of interest challenge existing schemes based on background subtraction (background motion too complex), multi-body motion estimation (insufficient parallax), and occlusion detection (uniformly textured background regions). We adopt a robust statistical inference approach to simultaneously estimate a maximally reduced regressor, and select regions that violate the null hypothesis (co-visibility under an epipolar domain deformation) as “salient”. We show that our algorithm can perform even in the absence of camera calibration information: while the resulting motion estimates would be incorrect, the partition of the domain into salient vs. non-salient is unaffected. We demonstrate our algorithm on video footage from helicopters, airplanes, and ground vehicles.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125014613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Aligning images in the wild 在野外对齐图像
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6247651
Wen-Yan Lin, Linlin Liu, Y. Matsushita, Kok-Lim Low, Siying Liu
{"title":"Aligning images in the wild","authors":"Wen-Yan Lin, Linlin Liu, Y. Matsushita, Kok-Lim Low, Siying Liu","doi":"10.1109/CVPR.2012.6247651","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6247651","url":null,"abstract":"Aligning image pairs with significant appearance change is a long standing computer vision challenge. Much of this problem stems from the local patch descriptors' instability to appearance variation. In this paper we suggest this instability is due less to descriptor corruption and more the difficulty in utilizing local information to canonically define the orientation (scale and rotation) at which a patch's descriptor should be computed. We address this issue by jointly estimating correspondence and relative patch orientation, within a hierarchical algorithm that utilizes a smoothly varying parameterization of geometric transformations. By collectively estimating the correspondence and orientation of all the features, we can align and orient features that cannot be stably matched with only local information. At the price of smoothing over motion discontinuities (due to independent motion or parallax), this approach can align image pairs that display significant inter-image appearance variations.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126092765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
A closed-form solution to uncalibrated photometric stereo via diffuse maxima 一个封闭形式的解决方案,以未经校准的光度立体通过漫射最大值
2012 IEEE Conference on Computer Vision and Pattern Recognition Pub Date : 2012-06-16 DOI: 10.1109/CVPR.2012.6247754
P. Favaro, Thoma Papadhimitri
{"title":"A closed-form solution to uncalibrated photometric stereo via diffuse maxima","authors":"P. Favaro, Thoma Papadhimitri","doi":"10.1109/CVPR.2012.6247754","DOIUrl":"https://doi.org/10.1109/CVPR.2012.6247754","url":null,"abstract":"In this paper we propose a novel solution to uncalibrated photometric stereo. Our approach is to eliminate the so-called generalized bas relief (GBR) ambiguity by exploiting points where the Lambertian reflection is maximal. We demonstrate several noteworthy properties of these maxima: 1) Closed-form solution: A single diffuse maximum constrains the GBR ambiguity to a semi-circle in 3D space; 2) Efficiency: As few as two diffuse maxima in different images identify a unique solution; 3) GBR-invariance: The estimation error of the GBR parameters is completely independent of the true parameters. Furthermore, our algorithm is remarkably robust: It can obtain an accurate estimate of the GBR parameters even with extremely high levels of outliers in the detected maxima (up to 80% of the observations). The method is validated on real data and achieves state-of-the-art results.","PeriodicalId":177454,"journal":{"name":"2012 IEEE Conference on Computer Vision and Pattern Recognition","volume":"430 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126093526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 61
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信