CVPR 2011最新文献

筛选
英文 中文
Iterative quantization: A procrustean approach to learning binary codes 迭代量化:一种学习二进制代码的普洛克斯坦方法
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995432
Yunchao Gong, S. Lazebnik
{"title":"Iterative quantization: A procrustean approach to learning binary codes","authors":"Yunchao Gong, S. Lazebnik","doi":"10.1109/CVPR.2011.5995432","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995432","url":null,"abstract":"This paper addresses the problem of learning similarity-preserving binary codes for efficient retrieval in large-scale image collections. We propose a simple and efficient alternating minimization scheme for finding a rotation of zero-centered data so as to minimize the quantization error of mapping this data to the vertices of a zero-centered binary hypercube. This method, dubbed iterative quantization (ITQ), has connections to multi-class spectral clustering and to the orthogonal Procrustes problem, and it can be used both with unsupervised data embeddings such as PCA and supervised embeddings such as canonical correlation analysis (CCA). Our experiments show that the resulting binary coding schemes decisively outperform several other state-of-the-art methods.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115349206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1189
A brute-force algorithm for reconstructing a scene from two projections 一种从两个投影重建场景的蛮力算法
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995669
O. Enqvist, Fangyuan Jiang, Fredrik Kahl
{"title":"A brute-force algorithm for reconstructing a scene from two projections","authors":"O. Enqvist, Fangyuan Jiang, Fredrik Kahl","doi":"10.1109/CVPR.2011.5995669","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995669","url":null,"abstract":"Is the real problem in finding the relative orientation of two viewpoints the correspondence problem? We argue that this is only one difficulty. Even with known correspondences, popular methods like the eight point algorithm and minimal solvers may break down due to planar scenes or small relative motions. In this paper, we derive a simple, brute-force algorithm which is both robust to outliers and has no such algorithmic degeneracies. Several cost functions are explored including maximizing the consensus set and robust norms like truncated least-squares. Our method is based on parameter search in a four-dimensional space using a new epipolar parametrization. In principle, we do an exhaustive search of parameter space, but the computations are very simple and easily parallelizable, resulting in an efficient method. Further speed-ups can be obtained by restricting the domain of possible motions to, for example, planar motions or small rotations. Experimental results are given for a variety of scenarios including scenes with a large portion of outliers. Further, we apply our algorithm to 3D motion segmentation where we outperform state-of-the-art on the well-known Hopkins-155 benchmark database.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115474490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Towards cross-category knowledge propagation for learning visual concepts 面向视觉概念学习的跨范畴知识传播
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995312
Guo-Jun Qi, C. Aggarwal, Y. Rui, Q. Tian, Shiyu Chang, Thomas S. Huang
{"title":"Towards cross-category knowledge propagation for learning visual concepts","authors":"Guo-Jun Qi, C. Aggarwal, Y. Rui, Q. Tian, Shiyu Chang, Thomas S. Huang","doi":"10.1109/CVPR.2011.5995312","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995312","url":null,"abstract":"In recent years, knowledge transfer algorithms have become one of most the active research areas in learning visual concepts. Most of the existing learning algorithms focuses on leveraging the knowledge transfer process which is specific to a given category. However, in many cases, such a process may not be very effective when a particular target category has very few samples. In such cases, it is interesting to examine, whether it is feasible to use cross-category knowledge for improving the learning process by exploring the knowledge in correlated categories. Such a task can be quite challenging due to variations in semantic similarities and differences between categories, which could either help or hinder the cross-category learning process. In order to address this challenge, we develop a cross-category label propagation algorithm, which can directly propagate the inter-category knowledge at instance level between the source and the target categories. Furthermore, this algorithm can automatically detect conditions under which the transfer process can be detrimental to the learning process. This provides us a way to know when the transfer of cross-category knowledge is both useful and desirable. We present experimental results on real image and video data sets in order to demonstrate the effectiveness of our approach.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"30 5","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120994205","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 81
Functional categorization of objects using real-time markerless motion capture 使用实时无标记动作捕捉的物体功能分类
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995582
Juergen Gall, A. Fossati, L. Gool
{"title":"Functional categorization of objects using real-time markerless motion capture","authors":"Juergen Gall, A. Fossati, L. Gool","doi":"10.1109/CVPR.2011.5995582","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995582","url":null,"abstract":"Unsupervised categorization of objects is a fundamental problem in computer vision. While appearance-based methods have become popular recently, other important cues like functionality are largely neglected. Motivated by psychological studies giving evidence that human demonstration has a facilitative effect on categorization in infancy, we propose an approach for object categorization from depth video streams. To this end, we have developed a method for capturing human motion in real-time. The captured data is then used to temporally segment the depth streams into actions. The set of segmented actions are then categorized in an un-supervised manner, through a novel descriptor for motion capture data that is robust to subject variations. Furthermore, we automatically localize the object that is manipulated within a video segment, and categorize it using the corresponding action. For evaluation, we have recorded a dataset that comprises depth data with registered video sequences for 6 subjects, 13 action classes, and 174 object manipulations.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121208182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 70
Local isomorphism to solve the pre-image problem in kernel methods 局部同构解决核方法中的预像问题
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995685
Dong Huang, Yuandong Tian, F. D. L. Torre
{"title":"Local isomorphism to solve the pre-image problem in kernel methods","authors":"Dong Huang, Yuandong Tian, F. D. L. Torre","doi":"10.1109/CVPR.2011.5995685","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995685","url":null,"abstract":"Kernel methods have been popular over the last decade to solve many computer vision, statistics and machine learning problems. An important, both theoretically and practically, open problem in kernel methods is the pre-image problem. The pre-image problem consists of finding a vector in the input space whose mapping is known in the feature space induced by a kernel. To solve the pre-image problem, this paper proposes a framework that computes an isomorphism between local Gram matrices in the input and feature space. Unlike existing methods that rely on analytic properties of kernels, our framework derives closed-form solutions to the pre-image problem in the case of non-differentiable and application-specific kernels. Experiments on the pre-image problem for visualizing cluster centers computed by kernel k-means and denoising high-dimensional images show that our algorithm outperforms state-of-the-art methods.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127502255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
An effective document image deblurring algorithm 一种有效的文档图像去模糊算法
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995568
Xiaogang Chen, Xiangjian He, Jie Yang, Qiang Wu
{"title":"An effective document image deblurring algorithm","authors":"Xiaogang Chen, Xiangjian He, Jie Yang, Qiang Wu","doi":"10.1109/CVPR.2011.5995568","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995568","url":null,"abstract":"Deblurring camera-based document image is an important task in digital document processing, since it can improve both the accuracy of optical character recognition systems and the visual quality of document images. Traditional deblurring algorithms have been proposed to work for natural-scene images. However the natural-scene images are not consistent with document images. In this paper, the distinct characteristics of document images are investigated. We propose a content-aware prior for document image deblurring. It is based on document image foreground segmentation. Besides, an upper-bound constraint combined with total variation based method is proposed to suppress the rings in the deblurred image. Comparing with the traditional general purpose deblurring methods, the proposed deblurring algorithm can produce more pleasing results on document images. Encouraging experimental results demonstrate the efficacy of the proposed method.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124825126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 84
Tracking 3D human pose with large root node uncertainty 具有大根节点不确定性的三维人体姿态跟踪
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995502
B. Daubney, Xianghua Xie
{"title":"Tracking 3D human pose with large root node uncertainty","authors":"B. Daubney, Xianghua Xie","doi":"10.1109/CVPR.2011.5995502","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995502","url":null,"abstract":"Representing articulated objects as a graphical model has gained much popularity in recent years, often the root node of the graph describes the global position and orientation of the object. In this work a method is presented to robustly track 3D human pose by permitting greater uncertainty to be modeled over the root node than existing techniques allow. Significantly, this is achieved without increasing the uncertainty of remaining parts of the model. The benefit is that a greater volume of the posterior can be supported making the approach less vulnerable to tracking failure. Given a hypothesis of the root node state a novel method is presented to estimate the posterior over the remaining parts of the body conditioned on this value. All probability distributions are approximated using a single Gaussian allowing inference to be carried out in closed form. A set of deterministically selected sample points are used that allow the posterior to be updated for each part requiring just seven image likelihood evaluations making it extremely efficient. Multiple root node states are supported and propagated using standard sampling techniques. We believe this to be the first work devoted to efficient tracking of human pose whilst modeling large uncertainty in the root node and demonstrate the presented method to be more robust to tracking failures than existing approaches.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124968844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
From partial shape matching through local deformation to robust global shape similarity for object detection 从局部变形的局部形状匹配到鲁棒的全局形状相似目标检测
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995591
Tianyang Ma, Longin Jan Latecki
{"title":"From partial shape matching through local deformation to robust global shape similarity for object detection","authors":"Tianyang Ma, Longin Jan Latecki","doi":"10.1109/CVPR.2011.5995591","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995591","url":null,"abstract":"In this paper, we propose a novel framework for contour based object detection. Compared to previous work, our contribution is three-fold. 1) A novel shape matching scheme suitable for partial matching of edge fragments. The shape descriptor has the same geometric units as shape context but our shape representation is not histogram based. 2) Grouping of partial matching hypotheses to object detection hypotheses is expressed as maximum clique inference on a weighted graph. 3) A novel local affine-transformation to utilize the holistic shape information for scoring and ranking the shape similarity hypotheses. Consequently, each detection result not only identifies the location of the target object in the image, but also provides a precise location of its contours, since we transform a complete model contour to the image. Very competitive results on ETHZ dataset, obtained in a pure shape-based framework, demonstrate that our method achieves not only accurate object detection but also precise contour localization on cluttered background.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126041884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 106
Spatial-DiscLDA for visual recognition 用于视觉识别的空间disclda
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995426
Zhenxing Niu, G. Hua, Xinbo Gao, Q. Tian
{"title":"Spatial-DiscLDA for visual recognition","authors":"Zhenxing Niu, G. Hua, Xinbo Gao, Q. Tian","doi":"10.1109/CVPR.2011.5995426","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995426","url":null,"abstract":"Topic models such as pLSA, LDA and their variants have been widely adopted for visual recognition. However, most of the adopted models, if not all, are unsupervised, which neglected the valuable supervised labels during model training. In this paper, we exploit recent advancement in supervised topic modeling, more particularly, the DiscLDA model for object recognition. We extend it to a part based visual representation to automatically identify and model different object parts. We call the proposed model as Spatial-DiscLDA (S-DiscLDA). It models the appearances and locations of the object parts simultaneously, which also takes the supervised labels into consideration. It can be directly used as a classifier to recognize the object. This is performed by an approximate inference algorithm based on Gibbs sampling and bridge sampling methods. We examine the performance of our model by comparing its performance with another supervised topic model on two scene category datasets, i.e., LabelMe and UIUC-sport dataset. We also compare our approach with other approaches which model spatial structures of visual features on the popular Caltech-4 dataset. The experimental results illustrate that it provides competitive performance.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123543456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Rank-SIFT: Learning to rank repeatable local interest points rank - sift:学习对可重复的局部兴趣点进行排序
CVPR 2011 Pub Date : 2011-06-20 DOI: 10.1109/CVPR.2011.5995461
Bing Li, Rong Xiao, Zhiwei Li, Rui Cai, Bao-Liang Lu, Lei Zhang
{"title":"Rank-SIFT: Learning to rank repeatable local interest points","authors":"Bing Li, Rong Xiao, Zhiwei Li, Rui Cai, Bao-Liang Lu, Lei Zhang","doi":"10.1109/CVPR.2011.5995461","DOIUrl":"https://doi.org/10.1109/CVPR.2011.5995461","url":null,"abstract":"Scale-invariant feature transform (SIFT) has been well studied in recent years. Most related research efforts focused on designing and learning effective descriptors to characterize a local interest point. However, how to identify stable local interest points is still a very challenging problem. In this paper, we propose a set of differential features, and based on them we adopt a data-driven approach to learn a ranking function to sort local interest points according to their stabilities across images containing the same visual objects. Compared with the handcrafted rule-based method used by the standard SIFT algorithm, our algorithm substantially improves the stability of detected local interest point on a very challenging benchmark dataset, in which images were generated under very different imaging conditions. Experimental results on the Oxford and PASCAL databases further demonstrate the superior performance of the proposed algorithm on both object image retrieval and category recognition.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125297335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 47
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信