从文本查询到视觉查询

N. Zikos, A. Delopoulos, Dafni Maria Vasilikari
{"title":"从文本查询到视觉查询","authors":"N. Zikos, A. Delopoulos, Dafni Maria Vasilikari","doi":"10.1109/CBMI.2016.7500270","DOIUrl":null,"url":null,"abstract":"In this paper we present a framework to transform textual queries into visual ones. The proposed method uses standard image retrieval techniques with textual queries and the Fast Geometric Consistency Test (FGCT) method. For every textual query a set of images is retrieved and for every image a set of descriptors is extracted. Extracted features are combined with respect to their similarity in their descriptors' space and afterwards with respect to their geometric consistency on the image plane. All pairs of images are tested for consistent geometric structures using the FGCT method. This procedure extracts the subset of images that have a persistent geometric formation in the descriptors' space. Descriptors that compose the persistent formation are extracted and used as the input in a visual query; those features constitute the visual context of the visual query. Afterwards we perform again the FGCT method, but this time using the set of extracted features of the persistent formation into the cloud of images that consists of images with out a priori textual knowledge. It is noteworthy that the proposed method is scale, rotation and translation invariant. Experimental results on the Microsoft's Clickture dataset which consist of 1 million images are presented to support these statements.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"From textual queries to visual queries\",\"authors\":\"N. Zikos, A. Delopoulos, Dafni Maria Vasilikari\",\"doi\":\"10.1109/CBMI.2016.7500270\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper we present a framework to transform textual queries into visual ones. The proposed method uses standard image retrieval techniques with textual queries and the Fast Geometric Consistency Test (FGCT) method. For every textual query a set of images is retrieved and for every image a set of descriptors is extracted. Extracted features are combined with respect to their similarity in their descriptors' space and afterwards with respect to their geometric consistency on the image plane. All pairs of images are tested for consistent geometric structures using the FGCT method. This procedure extracts the subset of images that have a persistent geometric formation in the descriptors' space. Descriptors that compose the persistent formation are extracted and used as the input in a visual query; those features constitute the visual context of the visual query. Afterwards we perform again the FGCT method, but this time using the set of extracted features of the persistent formation into the cloud of images that consists of images with out a priori textual knowledge. It is noteworthy that the proposed method is scale, rotation and translation invariant. Experimental results on the Microsoft's Clickture dataset which consist of 1 million images are presented to support these statements.\",\"PeriodicalId\":356608,\"journal\":{\"name\":\"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)\",\"volume\":\"11 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-06-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CBMI.2016.7500270\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CBMI.2016.7500270","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

本文提出了一个将文本查询转换为视觉查询的框架。该方法采用标准的图像检索技术和快速几何一致性测试(FGCT)方法。对于每个文本查询,检索一组图像,并为每个图像提取一组描述符。提取的特征根据其在描述符空间中的相似性进行组合,然后根据其在图像平面上的几何一致性进行组合。使用FGCT方法测试所有对图像的一致几何结构。该过程提取在描述符空间中具有持久几何形状的图像子集。组成持久形式的描述符被提取并用作可视化查询中的输入;这些特征构成了可视化查询的可视化上下文。之后,我们再次执行FGCT方法,但这次使用的是一组提取的持久形成的特征到图像云中,该图像由没有先验文本知识的图像组成。值得注意的是,该方法具有尺度、旋转、平移不变性。在微软的Clickture数据集(包含100万张图像)上的实验结果支持了这些说法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
From textual queries to visual queries
In this paper we present a framework to transform textual queries into visual ones. The proposed method uses standard image retrieval techniques with textual queries and the Fast Geometric Consistency Test (FGCT) method. For every textual query a set of images is retrieved and for every image a set of descriptors is extracted. Extracted features are combined with respect to their similarity in their descriptors' space and afterwards with respect to their geometric consistency on the image plane. All pairs of images are tested for consistent geometric structures using the FGCT method. This procedure extracts the subset of images that have a persistent geometric formation in the descriptors' space. Descriptors that compose the persistent formation are extracted and used as the input in a visual query; those features constitute the visual context of the visual query. Afterwards we perform again the FGCT method, but this time using the set of extracted features of the persistent formation into the cloud of images that consists of images with out a priori textual knowledge. It is noteworthy that the proposed method is scale, rotation and translation invariant. Experimental results on the Microsoft's Clickture dataset which consist of 1 million images are presented to support these statements.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信