{"title":"一对一的班级不平衡学习","authors":"Bilal Mirza, Zhiping Lin","doi":"10.1109/ICICS.2013.6782785","DOIUrl":null,"url":null,"abstract":"The performance of support vector machines (SVMs) can deteriorate when the number of samples in one class is much greater than that in the other. Existing methods tackle this problem by modifying the learning algorithms or resampling the datasets. In this paper, we propose a new method called one-vs-all for class imbalance learning (OVACIL) which neither modifies the SVM learning algorithms nor resamples the datasets. In the OVACIL method, we re-group a given imbalanced dataset into a number of new datasets comprising of all the original samples and train standard SVM classifiers using each of the datasets. The output scores of these classifiers on a testing sample are then compared and a final decision is made without a fixed decision threshold. This comparison is not biased toward any particular class, resulting in high accuracies of both classes. The Gmean and Fmeasure values obtained by OVACIL on 18 real-world imbalanced datasets surpass the previous best values reported by other state-of-the-art CIL methods on most of these datasets.","PeriodicalId":184544,"journal":{"name":"2013 9th International Conference on Information, Communications & Signal Processing","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"One-vs-all for class imbalance learning\",\"authors\":\"Bilal Mirza, Zhiping Lin\",\"doi\":\"10.1109/ICICS.2013.6782785\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The performance of support vector machines (SVMs) can deteriorate when the number of samples in one class is much greater than that in the other. Existing methods tackle this problem by modifying the learning algorithms or resampling the datasets. In this paper, we propose a new method called one-vs-all for class imbalance learning (OVACIL) which neither modifies the SVM learning algorithms nor resamples the datasets. In the OVACIL method, we re-group a given imbalanced dataset into a number of new datasets comprising of all the original samples and train standard SVM classifiers using each of the datasets. The output scores of these classifiers on a testing sample are then compared and a final decision is made without a fixed decision threshold. This comparison is not biased toward any particular class, resulting in high accuracies of both classes. The Gmean and Fmeasure values obtained by OVACIL on 18 real-world imbalanced datasets surpass the previous best values reported by other state-of-the-art CIL methods on most of these datasets.\",\"PeriodicalId\":184544,\"journal\":{\"name\":\"2013 9th International Conference on Information, Communications & Signal Processing\",\"volume\":\"8 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2013-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2013 9th International Conference on Information, Communications & Signal Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICICS.2013.6782785\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 9th International Conference on Information, Communications & Signal Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICICS.2013.6782785","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The performance of support vector machines (SVMs) can deteriorate when the number of samples in one class is much greater than that in the other. Existing methods tackle this problem by modifying the learning algorithms or resampling the datasets. In this paper, we propose a new method called one-vs-all for class imbalance learning (OVACIL) which neither modifies the SVM learning algorithms nor resamples the datasets. In the OVACIL method, we re-group a given imbalanced dataset into a number of new datasets comprising of all the original samples and train standard SVM classifiers using each of the datasets. The output scores of these classifiers on a testing sample are then compared and a final decision is made without a fixed decision threshold. This comparison is not biased toward any particular class, resulting in high accuracies of both classes. The Gmean and Fmeasure values obtained by OVACIL on 18 real-world imbalanced datasets surpass the previous best values reported by other state-of-the-art CIL methods on most of these datasets.