利用图像改进产品分类

A. Kannan, P. Talukdar, Nikhil Rasiwasia, Qifa Ke
{"title":"利用图像改进产品分类","authors":"A. Kannan, P. Talukdar, Nikhil Rasiwasia, Qifa Ke","doi":"10.1109/ICDM.2011.79","DOIUrl":null,"url":null,"abstract":"Product classification in Commerce search (\\eg{} Google Product Search, Bing Shopping) involves associating categories to offers of products from a large number of merchants. The categorized offers are used in many tasks including product taxonomy browsing and matching merchant offers to products in the catalog. Hence, learning a product classifier with high precision and recall is of fundamental importance in order to provide high quality shopping experience. A product offer typically consists of a short textual description and an image depicting the product. Traditional approaches to this classification task is to learn a classifier using only the textual descriptions of the products. In this paper, we show that the use of images, a weaker signal in our setting, in conjunction with the textual descriptions, a more discriminative signal, can considerably improve the precision of the classification task, irrespective of the type of classifier being used. We present a novel classification approach, \\Cross Adapt{} (\\CrossAdaptAcro{}), that is cognizant of the disparity in the discriminative power of different types of signals and hence makes use of the confusion matrix of dominant signal (text in our setting) to prudently leverage the weaker signal (image), for an improved performance. Our evaluation performed on data from a major Commerce search engine's catalog shows a 12\\% (absolute) improvement in precision at 100\\% coverage, and a 16\\% (absolute) improvement in recall at 90\\% precision compared to classifiers that only use textual description of products. In addition, \\CrossAdaptAcro{} also provides a more accurate classifier based only on the dominant signal (text) that can be used in situations in which only the dominant signal is available during application time.","PeriodicalId":106216,"journal":{"name":"2011 IEEE 11th International Conference on Data Mining","volume":"33 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"29","resultStr":"{\"title\":\"Improving Product Classification Using Images\",\"authors\":\"A. Kannan, P. Talukdar, Nikhil Rasiwasia, Qifa Ke\",\"doi\":\"10.1109/ICDM.2011.79\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Product classification in Commerce search (\\\\eg{} Google Product Search, Bing Shopping) involves associating categories to offers of products from a large number of merchants. The categorized offers are used in many tasks including product taxonomy browsing and matching merchant offers to products in the catalog. Hence, learning a product classifier with high precision and recall is of fundamental importance in order to provide high quality shopping experience. A product offer typically consists of a short textual description and an image depicting the product. Traditional approaches to this classification task is to learn a classifier using only the textual descriptions of the products. In this paper, we show that the use of images, a weaker signal in our setting, in conjunction with the textual descriptions, a more discriminative signal, can considerably improve the precision of the classification task, irrespective of the type of classifier being used. We present a novel classification approach, \\\\Cross Adapt{} (\\\\CrossAdaptAcro{}), that is cognizant of the disparity in the discriminative power of different types of signals and hence makes use of the confusion matrix of dominant signal (text in our setting) to prudently leverage the weaker signal (image), for an improved performance. Our evaluation performed on data from a major Commerce search engine's catalog shows a 12\\\\% (absolute) improvement in precision at 100\\\\% coverage, and a 16\\\\% (absolute) improvement in recall at 90\\\\% precision compared to classifiers that only use textual description of products. In addition, \\\\CrossAdaptAcro{} also provides a more accurate classifier based only on the dominant signal (text) that can be used in situations in which only the dominant signal is available during application time.\",\"PeriodicalId\":106216,\"journal\":{\"name\":\"2011 IEEE 11th International Conference on Data Mining\",\"volume\":\"33 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2011-12-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"29\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2011 IEEE 11th International Conference on Data Mining\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICDM.2011.79\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 IEEE 11th International Conference on Data Mining","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDM.2011.79","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 29

摘要

商业搜索中的产品分类(\eg{}谷歌Product search, Bing Shopping)涉及将类别与来自大量商家的产品报价相关联。分类优惠用于许多任务,包括产品分类浏览和将商家优惠与目录中的产品相匹配。因此,学习一个具有高准确率和召回率的产品分类器对于提供高质量的购物体验至关重要。产品报价通常由简短的文本描述和描述产品的图像组成。该分类任务的传统方法是仅使用产品的文本描述来学习分类器。在本文中,我们展示了在我们的设置中使用图像,一个较弱的信号,结合文本描述,一个更具判别性的信号,可以大大提高分类任务的精度,而不管使用哪种分类器。我们提出了一种新的分类方法,\CrossAdapt {} (\CrossAdaptAcro{}),它认识到不同类型信号的鉴别能力的差异,因此利用主导信号(我们设置中的文本)的混淆矩阵来谨慎地利用较弱的信号(图像),以提高性能。我们对来自主要商业搜索引擎目录的数据进行的评估显示,与仅使用产品文本描述的分类器相比,在100%覆盖率下,准确率提高了12%(绝对),在90%准确率下,召回率提高了16%(绝对)。此外,\CrossAdaptAcro{}还提供了一个仅基于主导信号(文本)的更准确的分类器,可用于在应用程序期间只有主导信号可用的情况。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Improving Product Classification Using Images
Product classification in Commerce search (\eg{} Google Product Search, Bing Shopping) involves associating categories to offers of products from a large number of merchants. The categorized offers are used in many tasks including product taxonomy browsing and matching merchant offers to products in the catalog. Hence, learning a product classifier with high precision and recall is of fundamental importance in order to provide high quality shopping experience. A product offer typically consists of a short textual description and an image depicting the product. Traditional approaches to this classification task is to learn a classifier using only the textual descriptions of the products. In this paper, we show that the use of images, a weaker signal in our setting, in conjunction with the textual descriptions, a more discriminative signal, can considerably improve the precision of the classification task, irrespective of the type of classifier being used. We present a novel classification approach, \Cross Adapt{} (\CrossAdaptAcro{}), that is cognizant of the disparity in the discriminative power of different types of signals and hence makes use of the confusion matrix of dominant signal (text in our setting) to prudently leverage the weaker signal (image), for an improved performance. Our evaluation performed on data from a major Commerce search engine's catalog shows a 12\% (absolute) improvement in precision at 100\% coverage, and a 16\% (absolute) improvement in recall at 90\% precision compared to classifiers that only use textual description of products. In addition, \CrossAdaptAcro{} also provides a more accurate classifier based only on the dominant signal (text) that can be used in situations in which only the dominant signal is available during application time.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信