{"title":"基于本地化内容的图像检索与自学多实例学习","authors":"Qifeng Qiao, P. Beling","doi":"10.1109/ICDMW.2009.105","DOIUrl":null,"url":null,"abstract":"There are many scenarios in which multi-instance learning problems may be difficult to solve because of a lack of correctly labeled examples for algorithm training. Labeled examples may be difficult or expensive to obtain because human effort is often needed to produce labels and because there may be limitations on the ability to collect large samples for training from a homogeneous population. In this paper, we present a technique called self-taught multiple-instance learning (STMIL) that deals with learning from a limited number of ambiguously labeled examples. STMIL uses a sparse representation for examples belonging to different classes in terms of a shared dictionary derived from the unlabeled data. This sparse representation can be optimized under the multiple instance setting to both construct high-level features and unite the data distribution. We present an optimization procedure for STMIL along with experiments on localized content-based image retrieval. Our experimental results suggest that, though it learns from a small number of labeled examples, STMIL is superior to standard algorithms in terms of computational efficiency and is at least competitive in terms of accuracy.","PeriodicalId":351078,"journal":{"name":"2009 IEEE International Conference on Data Mining Workshops","volume":"226 6","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2009-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Localized Content Based Image Retrieval with Self-Taught Multiple Instance Learning\",\"authors\":\"Qifeng Qiao, P. Beling\",\"doi\":\"10.1109/ICDMW.2009.105\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"There are many scenarios in which multi-instance learning problems may be difficult to solve because of a lack of correctly labeled examples for algorithm training. Labeled examples may be difficult or expensive to obtain because human effort is often needed to produce labels and because there may be limitations on the ability to collect large samples for training from a homogeneous population. In this paper, we present a technique called self-taught multiple-instance learning (STMIL) that deals with learning from a limited number of ambiguously labeled examples. STMIL uses a sparse representation for examples belonging to different classes in terms of a shared dictionary derived from the unlabeled data. This sparse representation can be optimized under the multiple instance setting to both construct high-level features and unite the data distribution. We present an optimization procedure for STMIL along with experiments on localized content-based image retrieval. Our experimental results suggest that, though it learns from a small number of labeled examples, STMIL is superior to standard algorithms in terms of computational efficiency and is at least competitive in terms of accuracy.\",\"PeriodicalId\":351078,\"journal\":{\"name\":\"2009 IEEE International Conference on Data Mining Workshops\",\"volume\":\"226 6\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2009-12-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2009 IEEE International Conference on Data Mining Workshops\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICDMW.2009.105\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2009 IEEE International Conference on Data Mining Workshops","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDMW.2009.105","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Localized Content Based Image Retrieval with Self-Taught Multiple Instance Learning
There are many scenarios in which multi-instance learning problems may be difficult to solve because of a lack of correctly labeled examples for algorithm training. Labeled examples may be difficult or expensive to obtain because human effort is often needed to produce labels and because there may be limitations on the ability to collect large samples for training from a homogeneous population. In this paper, we present a technique called self-taught multiple-instance learning (STMIL) that deals with learning from a limited number of ambiguously labeled examples. STMIL uses a sparse representation for examples belonging to different classes in terms of a shared dictionary derived from the unlabeled data. This sparse representation can be optimized under the multiple instance setting to both construct high-level features and unite the data distribution. We present an optimization procedure for STMIL along with experiments on localized content-based image retrieval. Our experimental results suggest that, though it learns from a small number of labeled examples, STMIL is superior to standard algorithms in terms of computational efficiency and is at least competitive in terms of accuracy.