Akihiko Nakagawa, Andrea Kutics, Kiyotaka Tanaka, Masaomi Nakajima
{"title":"结合词和基于对象的视觉特征在图像检索中的应用","authors":"Akihiko Nakagawa, Andrea Kutics, Kiyotaka Tanaka, Masaomi Nakajima","doi":"10.1109/ICIAP.2003.1234075","DOIUrl":null,"url":null,"abstract":"The paper presents a novel approach for image retrieval by combining textual and object-based visual features in order to reduce the inconsistency between the subjective user's similarity interpretation and the retrieval results produced by objective similarity models. A novel multi-scale segmentation framework is proposed to detect prominent image objects. These objects are clustered according to their visual features and mapped to related words determined by psychophysical studies. Furthermore, a hierarchy of words expressing higher-level meaning is determined on the basis of natural language processing and user evaluation. Experiments conducted on a large set of natural images showed that higher retrieval precision in terms of estimating user retrieval semantics could be achieved via this two-layer word association and also by supporting various query specifications and options.","PeriodicalId":218076,"journal":{"name":"12th International Conference on Image Analysis and Processing, 2003.Proceedings.","volume":"26 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2003-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"Combining words and object-based visual features in image retrieval\",\"authors\":\"Akihiko Nakagawa, Andrea Kutics, Kiyotaka Tanaka, Masaomi Nakajima\",\"doi\":\"10.1109/ICIAP.2003.1234075\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The paper presents a novel approach for image retrieval by combining textual and object-based visual features in order to reduce the inconsistency between the subjective user's similarity interpretation and the retrieval results produced by objective similarity models. A novel multi-scale segmentation framework is proposed to detect prominent image objects. These objects are clustered according to their visual features and mapped to related words determined by psychophysical studies. Furthermore, a hierarchy of words expressing higher-level meaning is determined on the basis of natural language processing and user evaluation. Experiments conducted on a large set of natural images showed that higher retrieval precision in terms of estimating user retrieval semantics could be achieved via this two-layer word association and also by supporting various query specifications and options.\",\"PeriodicalId\":218076,\"journal\":{\"name\":\"12th International Conference on Image Analysis and Processing, 2003.Proceedings.\",\"volume\":\"26 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2003-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"12th International Conference on Image Analysis and Processing, 2003.Proceedings.\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICIAP.2003.1234075\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"12th International Conference on Image Analysis and Processing, 2003.Proceedings.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICIAP.2003.1234075","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Combining words and object-based visual features in image retrieval
The paper presents a novel approach for image retrieval by combining textual and object-based visual features in order to reduce the inconsistency between the subjective user's similarity interpretation and the retrieval results produced by objective similarity models. A novel multi-scale segmentation framework is proposed to detect prominent image objects. These objects are clustered according to their visual features and mapped to related words determined by psychophysical studies. Furthermore, a hierarchy of words expressing higher-level meaning is determined on the basis of natural language processing and user evaluation. Experiments conducted on a large set of natural images showed that higher retrieval precision in terms of estimating user retrieval semantics could be achieved via this two-layer word association and also by supporting various query specifications and options.