{"title":"2-DOM:面向Web图像标注的二维对象模型","authors":"S. Alcic, Stefan Conrad","doi":"10.1109/SMAP.2008.23","DOIUrl":null,"url":null,"abstract":"The automatic annotation of images is still a non-reliable process due to the well-known semantic gap dominating between the physical representation of images and their high level semantics. To avoid the confrontation with the semantic gap several approaches restrict the image dataset to web images. Web images mostly appear on websites with other text contents which can deliver important information about the image semantics. Popular image search engines use text contents surrounding the image to generate annotation keywords. Also emphasized text contents like headlines are assumed to be important description providers. Otherwise we discover false positive results in high ranking positions of this search engines which are the effect of incorrect text-to-image mappings. This paper addresses the problem of finding correct matches between text elements and images in HTML-documents by extending the DOM model tree of a web-page to a 2-dimensional object model (2-DOM) tree. This model adapts to the two dimensional manner of web documents and thus allows a better mapping of text articles to images. The evaluation results show that with a precision over 90 percent text articles are well assigned to web images. This leads intuitive to better textual information about the presented events on the images and thus yields to a better image retrieval quality in querying-by-keyword scenarios.","PeriodicalId":292389,"journal":{"name":"2008 Third International Workshop on Semantic Media Adaptation and Personalization","volume":"39 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"2-DOM: A 2-Dimensional Object Model towards Web Image Annotation\",\"authors\":\"S. Alcic, Stefan Conrad\",\"doi\":\"10.1109/SMAP.2008.23\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The automatic annotation of images is still a non-reliable process due to the well-known semantic gap dominating between the physical representation of images and their high level semantics. To avoid the confrontation with the semantic gap several approaches restrict the image dataset to web images. Web images mostly appear on websites with other text contents which can deliver important information about the image semantics. Popular image search engines use text contents surrounding the image to generate annotation keywords. Also emphasized text contents like headlines are assumed to be important description providers. Otherwise we discover false positive results in high ranking positions of this search engines which are the effect of incorrect text-to-image mappings. This paper addresses the problem of finding correct matches between text elements and images in HTML-documents by extending the DOM model tree of a web-page to a 2-dimensional object model (2-DOM) tree. This model adapts to the two dimensional manner of web documents and thus allows a better mapping of text articles to images. The evaluation results show that with a precision over 90 percent text articles are well assigned to web images. This leads intuitive to better textual information about the presented events on the images and thus yields to a better image retrieval quality in querying-by-keyword scenarios.\",\"PeriodicalId\":292389,\"journal\":{\"name\":\"2008 Third International Workshop on Semantic Media Adaptation and Personalization\",\"volume\":\"39 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2008-12-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2008 Third International Workshop on Semantic Media Adaptation and Personalization\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SMAP.2008.23\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 Third International Workshop on Semantic Media Adaptation and Personalization","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SMAP.2008.23","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
2-DOM: A 2-Dimensional Object Model towards Web Image Annotation
The automatic annotation of images is still a non-reliable process due to the well-known semantic gap dominating between the physical representation of images and their high level semantics. To avoid the confrontation with the semantic gap several approaches restrict the image dataset to web images. Web images mostly appear on websites with other text contents which can deliver important information about the image semantics. Popular image search engines use text contents surrounding the image to generate annotation keywords. Also emphasized text contents like headlines are assumed to be important description providers. Otherwise we discover false positive results in high ranking positions of this search engines which are the effect of incorrect text-to-image mappings. This paper addresses the problem of finding correct matches between text elements and images in HTML-documents by extending the DOM model tree of a web-page to a 2-dimensional object model (2-DOM) tree. This model adapts to the two dimensional manner of web documents and thus allows a better mapping of text articles to images. The evaluation results show that with a precision over 90 percent text articles are well assigned to web images. This leads intuitive to better textual information about the presented events on the images and thus yields to a better image retrieval quality in querying-by-keyword scenarios.