{"title":"基于核方法的信息粒优化和联合训练","authors":"Yuzhang Bai , Jusheng Mi , Leijun Li","doi":"10.1016/j.asoc.2024.111584","DOIUrl":null,"url":null,"abstract":"<div><p>Co-training was originally designed for multi-view data. Subsequent theoretical research has extended co-training to the application of single-view data. The construction of a feature subspace is one of the methods to expand single-view data into multi-view data, so the establishment of a feature subspace is the key to this approach. In this paper, the kernel method is used to form an implicit feature subspace, to perform multi-view co-training on single-view data. Because the attribute value in the feature subspace is not known, the subspace is a pseudo-view. To make the view adapt to the base classifier, this paper uses the neighborhood classifier as the base classifier and proposes an adaptive kernel function and three kernel parameter optimization methods according to the characteristics of the neighborhood classifier to build the feature subspace adapted to the neighborhood classifier. The decision of the neighborhood classifier needs to be made based on information granules generated by unlabeled objects. In the iteration of the feature subspace, we can continuously learn and optimize the information granule, and finally form what we expect, to get the implicit feature space corresponding to the granule and improve the accuracy of the base classifier. Finally, taking five data sets from UCI, and using accuracy and F1-score as evaluation indicators, we conduct a comparative experiment. The experimental results show that an adaptive kernel function, three kernel parameter optimization methods, and the co-training method presented in this paper are effective.</p></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":"158 ","pages":"Article 111584"},"PeriodicalIF":6.6000,"publicationDate":"2024-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Information granule optimization and co-training based on kernel method\",\"authors\":\"Yuzhang Bai , Jusheng Mi , Leijun Li\",\"doi\":\"10.1016/j.asoc.2024.111584\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Co-training was originally designed for multi-view data. Subsequent theoretical research has extended co-training to the application of single-view data. The construction of a feature subspace is one of the methods to expand single-view data into multi-view data, so the establishment of a feature subspace is the key to this approach. In this paper, the kernel method is used to form an implicit feature subspace, to perform multi-view co-training on single-view data. Because the attribute value in the feature subspace is not known, the subspace is a pseudo-view. To make the view adapt to the base classifier, this paper uses the neighborhood classifier as the base classifier and proposes an adaptive kernel function and three kernel parameter optimization methods according to the characteristics of the neighborhood classifier to build the feature subspace adapted to the neighborhood classifier. The decision of the neighborhood classifier needs to be made based on information granules generated by unlabeled objects. In the iteration of the feature subspace, we can continuously learn and optimize the information granule, and finally form what we expect, to get the implicit feature space corresponding to the granule and improve the accuracy of the base classifier. Finally, taking five data sets from UCI, and using accuracy and F1-score as evaluation indicators, we conduct a comparative experiment. The experimental results show that an adaptive kernel function, three kernel parameter optimization methods, and the co-training method presented in this paper are effective.</p></div>\",\"PeriodicalId\":50737,\"journal\":{\"name\":\"Applied Soft Computing\",\"volume\":\"158 \",\"pages\":\"Article 111584\"},\"PeriodicalIF\":6.6000,\"publicationDate\":\"2024-04-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Soft Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1568494624003582\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1568494624003582","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
摘要
联合训练最初是针对多视角数据设计的。后来的理论研究将协同训练扩展到单视角数据的应用。构建特征子空间是将单视角数据扩展为多视角数据的方法之一,因此特征子空间的建立是这种方法的关键。本文采用核方法形成隐式特征子空间,对单视角数据进行多视角协同训练。由于不知道特征子空间中的属性值,因此该子空间是一个伪视图。为了使视图适应基分类器,本文以邻域分类器为基分类器,并根据邻域分类器的特点提出了自适应核函数和三种核参数优化方法,以构建适应邻域分类器的特征子空间。邻域分类器的决策需要基于未标记对象产生的信息颗粒。在特征子空间的迭代中,我们可以不断学习和优化信息颗粒,最终形成我们所期望的,得到与颗粒相对应的隐含特征空间,提高基础分类器的准确率。最后,以 UCI 的五组数据为研究对象,以准确率和 F1 分数为评价指标,进行对比实验。实验结果表明,自适应内核函数、三种内核参数优化方法以及本文提出的协同训练方法都是有效的。
Information granule optimization and co-training based on kernel method
Co-training was originally designed for multi-view data. Subsequent theoretical research has extended co-training to the application of single-view data. The construction of a feature subspace is one of the methods to expand single-view data into multi-view data, so the establishment of a feature subspace is the key to this approach. In this paper, the kernel method is used to form an implicit feature subspace, to perform multi-view co-training on single-view data. Because the attribute value in the feature subspace is not known, the subspace is a pseudo-view. To make the view adapt to the base classifier, this paper uses the neighborhood classifier as the base classifier and proposes an adaptive kernel function and three kernel parameter optimization methods according to the characteristics of the neighborhood classifier to build the feature subspace adapted to the neighborhood classifier. The decision of the neighborhood classifier needs to be made based on information granules generated by unlabeled objects. In the iteration of the feature subspace, we can continuously learn and optimize the information granule, and finally form what we expect, to get the implicit feature space corresponding to the granule and improve the accuracy of the base classifier. Finally, taking five data sets from UCI, and using accuracy and F1-score as evaluation indicators, we conduct a comparative experiment. The experimental results show that an adaptive kernel function, three kernel parameter optimization methods, and the co-training method presented in this paper are effective.
期刊介绍:
Applied Soft Computing is an international journal promoting an integrated view of soft computing to solve real life problems.The focus is to publish the highest quality research in application and convergence of the areas of Fuzzy Logic, Neural Networks, Evolutionary Computing, Rough Sets and other similar techniques to address real world complexities.
Applied Soft Computing is a rolling publication: articles are published as soon as the editor-in-chief has accepted them. Therefore, the web site will continuously be updated with new articles and the publication time will be short.