{"title":"利用视频本体实现快速准确的逐例查询检索","authors":"Kimiaki Shirahama, K. Uehara","doi":"10.1109/ICSC.2011.88","DOIUrl":null,"url":null,"abstract":"In this paper, we develop a video retrieval method based on Query-By-Example (QBE) approach where example shots are provided to represent a query, and used to construct a retrieval model. One drawback of QBE is that a user can only provide a small number of example shots, while each shot is generally represented by a high-dimensional feature. This causes that the retrieval model tends to be over fit to feature dimensions which are specific to example shots, but are ineffective for retrieving relevant shots. As a result, many clearly irrelevant shots are retrieved. To overcome this, we construct a {\\it video ontology} as knowledge base for QBE. Our video ontology is used to select concepts related to a query. Then, irrelevant shots are filtered by referring to recognition results of objects corresponding to selected concepts. Also, counter-example shots are not provided in QBE, although they are useful for constructing an accurate retrieval model. We introduce a method which selects counter-example shots among shots without user supervision. In this method, our video ontology is used to exclude shots relevant to the query from candidates of counter-example shots. Specifically, we filter shots where object recognition results for concepts related to the query are similar to those of example shots. The effectiveness of our video ontology is tested on TRECVID 2009 video data.","PeriodicalId":408382,"journal":{"name":"2011 IEEE Fifth International Conference on Semantic Computing","volume":"124 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Utilizing Video Ontology for Fast and Accurate Query-by-Example Retrieval\",\"authors\":\"Kimiaki Shirahama, K. Uehara\",\"doi\":\"10.1109/ICSC.2011.88\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we develop a video retrieval method based on Query-By-Example (QBE) approach where example shots are provided to represent a query, and used to construct a retrieval model. One drawback of QBE is that a user can only provide a small number of example shots, while each shot is generally represented by a high-dimensional feature. This causes that the retrieval model tends to be over fit to feature dimensions which are specific to example shots, but are ineffective for retrieving relevant shots. As a result, many clearly irrelevant shots are retrieved. To overcome this, we construct a {\\\\it video ontology} as knowledge base for QBE. Our video ontology is used to select concepts related to a query. Then, irrelevant shots are filtered by referring to recognition results of objects corresponding to selected concepts. Also, counter-example shots are not provided in QBE, although they are useful for constructing an accurate retrieval model. We introduce a method which selects counter-example shots among shots without user supervision. In this method, our video ontology is used to exclude shots relevant to the query from candidates of counter-example shots. Specifically, we filter shots where object recognition results for concepts related to the query are similar to those of example shots. The effectiveness of our video ontology is tested on TRECVID 2009 video data.\",\"PeriodicalId\":408382,\"journal\":{\"name\":\"2011 IEEE Fifth International Conference on Semantic Computing\",\"volume\":\"124 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2011-09-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2011 IEEE Fifth International Conference on Semantic Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICSC.2011.88\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 IEEE Fifth International Conference on Semantic Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSC.2011.88","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
摘要
在本文中,我们开发了一种基于QBE (query - by - example)方法的视频检索方法,该方法通过提供示例镜头来表示查询,并用于构建检索模型。QBE的一个缺点是用户只能提供少量的示例镜头,而每个镜头通常由一个高维特征表示。这导致检索模型倾向于过度拟合特定于示例镜头的特征维度,但对于检索相关镜头无效。结果,许多明显不相关的镜头被检索出来。为了克服这个问题,我们构建了一个{\it视频本体}作为QBE的知识库。我们的视频本体用于选择与查询相关的概念。然后,根据所选概念对应对象的识别结果,过滤出不相关的镜头。此外,QBE中没有提供反例镜头,尽管它们对于构建准确的检索模型很有用。我们介绍了一种在没有用户监督的情况下,从镜头中选择反例镜头的方法。在该方法中,我们的视频本体被用来从反例镜头的候选中排除与查询相关的镜头。具体来说,我们对与查询相关的概念的对象识别结果与示例镜头相似的镜头进行过滤。在TRECVID 2009视频数据上测试了我们的视频本体的有效性。
Utilizing Video Ontology for Fast and Accurate Query-by-Example Retrieval
In this paper, we develop a video retrieval method based on Query-By-Example (QBE) approach where example shots are provided to represent a query, and used to construct a retrieval model. One drawback of QBE is that a user can only provide a small number of example shots, while each shot is generally represented by a high-dimensional feature. This causes that the retrieval model tends to be over fit to feature dimensions which are specific to example shots, but are ineffective for retrieving relevant shots. As a result, many clearly irrelevant shots are retrieved. To overcome this, we construct a {\it video ontology} as knowledge base for QBE. Our video ontology is used to select concepts related to a query. Then, irrelevant shots are filtered by referring to recognition results of objects corresponding to selected concepts. Also, counter-example shots are not provided in QBE, although they are useful for constructing an accurate retrieval model. We introduce a method which selects counter-example shots among shots without user supervision. In this method, our video ontology is used to exclude shots relevant to the query from candidates of counter-example shots. Specifically, we filter shots where object recognition results for concepts related to the query are similar to those of example shots. The effectiveness of our video ontology is tested on TRECVID 2009 video data.