Spatial-DiscLDA for visual recognition

Zhenxing Niu, G. Hua, Xinbo Gao, Q. Tian
{"title":"Spatial-DiscLDA for visual recognition","authors":"Zhenxing Niu, G. Hua, Xinbo Gao, Q. Tian","doi":"10.1109/CVPR.2011.5995426","DOIUrl":null,"url":null,"abstract":"Topic models such as pLSA, LDA and their variants have been widely adopted for visual recognition. However, most of the adopted models, if not all, are unsupervised, which neglected the valuable supervised labels during model training. In this paper, we exploit recent advancement in supervised topic modeling, more particularly, the DiscLDA model for object recognition. We extend it to a part based visual representation to automatically identify and model different object parts. We call the proposed model as Spatial-DiscLDA (S-DiscLDA). It models the appearances and locations of the object parts simultaneously, which also takes the supervised labels into consideration. It can be directly used as a classifier to recognize the object. This is performed by an approximate inference algorithm based on Gibbs sampling and bridge sampling methods. We examine the performance of our model by comparing its performance with another supervised topic model on two scene category datasets, i.e., LabelMe and UIUC-sport dataset. We also compare our approach with other approaches which model spatial structures of visual features on the popular Caltech-4 dataset. The experimental results illustrate that it provides competitive performance.","PeriodicalId":445398,"journal":{"name":"CVPR 2011","volume":"59 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"35","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"CVPR 2011","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CVPR.2011.5995426","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 35

Abstract

Topic models such as pLSA, LDA and their variants have been widely adopted for visual recognition. However, most of the adopted models, if not all, are unsupervised, which neglected the valuable supervised labels during model training. In this paper, we exploit recent advancement in supervised topic modeling, more particularly, the DiscLDA model for object recognition. We extend it to a part based visual representation to automatically identify and model different object parts. We call the proposed model as Spatial-DiscLDA (S-DiscLDA). It models the appearances and locations of the object parts simultaneously, which also takes the supervised labels into consideration. It can be directly used as a classifier to recognize the object. This is performed by an approximate inference algorithm based on Gibbs sampling and bridge sampling methods. We examine the performance of our model by comparing its performance with another supervised topic model on two scene category datasets, i.e., LabelMe and UIUC-sport dataset. We also compare our approach with other approaches which model spatial structures of visual features on the popular Caltech-4 dataset. The experimental results illustrate that it provides competitive performance.
用于视觉识别的空间disclda
主题模型如pLSA、LDA及其变体已被广泛应用于视觉识别。然而,大多数采用的模型(如果不是全部的话)都是无监督的,这在模型训练过程中忽略了有价值的监督标签。在本文中,我们利用监督主题建模的最新进展,特别是用于对象识别的DiscLDA模型。我们将其扩展为基于部件的可视化表示,以自动识别和建模不同的对象部件。我们将提出的模型称为空间- disclda (S-DiscLDA)。它同时对物体部件的外观和位置进行建模,并考虑了监督标签。它可以直接用作分类器来识别对象。这是通过基于Gibbs抽样和桥式抽样方法的近似推理算法来实现的。我们通过将我们的模型与另一个监督主题模型在两个场景类别数据集(即LabelMe和UIUC-sport数据集)上的性能进行比较,来检查我们的模型的性能。我们还将我们的方法与在流行的Caltech-4数据集上对视觉特征的空间结构建模的其他方法进行了比较。实验结果表明,该方法具有较好的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信