{"title":"基于解释的物体识别","authors":"Geoffrey Levine, G. DeJong","doi":"10.1109/WACV.2008.4544019","DOIUrl":null,"url":null,"abstract":"Many of today's visual scene and object categorization systems learn to classify using a statistical profile over a large number of small-scale local features sampled from the image. While some application systems have been constructed, this technology has enjoyed far more success in the research setting. The approach is best suited to tasks where within-class variability is small compared to between-class variability. This condition holds for large diverse artificial collections such as CalTech 101 where most categories have little to do with each other, but it often does not hold among naturalistic application-driven categories. Here, category distinctions are more likely to be conceptual or functional, and within-class differences can rival or exceed between- class differences. In this paper, we show how the local feature approach can be extended using explanation-based learning (EBL). The EBL approach makes use of readily available prior domain knowledge assembled into plausible explanations for why a training example's observable features might merit its assigned training label. Explanations expose additional semantic features and suggest how those hidden features may be estimated from observable features. We exhibit our approach on two CalTech 101 dataset tasks that we argue are emblematic of applied domains: Ketch vs. Schooner and Airplane vs. Background. In both cases classification accuracy is significantly improved.","PeriodicalId":439571,"journal":{"name":"2008 IEEE Workshop on Applications of Computer Vision","volume":"58 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Explanation-Based Object Recognition\",\"authors\":\"Geoffrey Levine, G. DeJong\",\"doi\":\"10.1109/WACV.2008.4544019\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Many of today's visual scene and object categorization systems learn to classify using a statistical profile over a large number of small-scale local features sampled from the image. While some application systems have been constructed, this technology has enjoyed far more success in the research setting. The approach is best suited to tasks where within-class variability is small compared to between-class variability. This condition holds for large diverse artificial collections such as CalTech 101 where most categories have little to do with each other, but it often does not hold among naturalistic application-driven categories. Here, category distinctions are more likely to be conceptual or functional, and within-class differences can rival or exceed between- class differences. In this paper, we show how the local feature approach can be extended using explanation-based learning (EBL). The EBL approach makes use of readily available prior domain knowledge assembled into plausible explanations for why a training example's observable features might merit its assigned training label. Explanations expose additional semantic features and suggest how those hidden features may be estimated from observable features. We exhibit our approach on two CalTech 101 dataset tasks that we argue are emblematic of applied domains: Ketch vs. Schooner and Airplane vs. Background. In both cases classification accuracy is significantly improved.\",\"PeriodicalId\":439571,\"journal\":{\"name\":\"2008 IEEE Workshop on Applications of Computer Vision\",\"volume\":\"58 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2008-01-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2008 IEEE Workshop on Applications of Computer Vision\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/WACV.2008.4544019\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 IEEE Workshop on Applications of Computer Vision","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WACV.2008.4544019","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
今天的许多视觉场景和对象分类系统学习使用从图像中采样的大量小尺度局部特征的统计概况进行分类。虽然已经构建了一些应用系统,但该技术在研究环境中取得了更大的成功。该方法最适合于类内可变性比类间可变性小的任务。这种情况适用于各种各样的大型人工集合,如CalTech 101,其中大多数类别彼此之间几乎没有关系,但在自然应用驱动的类别中往往不适用。在这里,类别差异更可能是概念性的或功能性的,类内差异可以与类间差异相媲美或超过。在本文中,我们展示了如何使用基于解释的学习(EBL)扩展局部特征方法。EBL方法利用现成的先验领域知识组装成合理的解释,来解释为什么训练示例的可观察特征可能值得分配其训练标签。解释揭示了额外的语义特征,并建议如何从可观察的特征中估计这些隐藏的特征。我们在两个CalTech 101数据集任务上展示了我们的方法,我们认为这两个任务是应用领域的象征:Ketch vs. Schooner和Airplane vs. Background。在这两种情况下,分类精度都得到了显著提高。
Many of today's visual scene and object categorization systems learn to classify using a statistical profile over a large number of small-scale local features sampled from the image. While some application systems have been constructed, this technology has enjoyed far more success in the research setting. The approach is best suited to tasks where within-class variability is small compared to between-class variability. This condition holds for large diverse artificial collections such as CalTech 101 where most categories have little to do with each other, but it often does not hold among naturalistic application-driven categories. Here, category distinctions are more likely to be conceptual or functional, and within-class differences can rival or exceed between- class differences. In this paper, we show how the local feature approach can be extended using explanation-based learning (EBL). The EBL approach makes use of readily available prior domain knowledge assembled into plausible explanations for why a training example's observable features might merit its assigned training label. Explanations expose additional semantic features and suggest how those hidden features may be estimated from observable features. We exhibit our approach on two CalTech 101 dataset tasks that we argue are emblematic of applied domains: Ketch vs. Schooner and Airplane vs. Background. In both cases classification accuracy is significantly improved.