{"title":"MIAC: Mutual-Information Classifier with ADASYN for Imbalanced Classification","authors":"Yanyu Cao, Xiaodong Zhao, Zhiping Zhou, Yufei Chen, Xianhui Liu, Yongming Lang","doi":"10.1109/SPAC46244.2018.8965597","DOIUrl":null,"url":null,"abstract":"currently, classification of imbalanced data is a significant issue in the area of data mining and machine learning because of the imbalance of most of the data set. An effective solution of this problem is Cost-Sensitive Learning (CSL), but when the costs are not given, this method cannot work property. As a Cost-Free Learning (CFL) method, Mutual-Information Classification (MIC) can obtain the optimal classification results when the cost information is not given. But this method emphasizes the data of minority class too much and neglects the accuracy of the classification of majority class. And based on the above, this paper presented a CFL method called Mutual-Information-ADASYN Classification (MIAC). Firstly, we get the abstaining samples which are hard to be classified by using MIC. Then we use these abstention samples to synthesize new instance by using the method of ADASYN. Thirdly, we build Mutual- Information-ADASYN Classification using the new samples. Finally, we use our classifier to get the final results. We evaluated the performance of MIAC on several imbalance binary datasets with different imbalance ratios. The experimental results indicate that the MIAC is more effective than MIC on dealing with imbalanced datasets.","PeriodicalId":360369,"journal":{"name":"2018 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SPAC46244.2018.8965597","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
currently, classification of imbalanced data is a significant issue in the area of data mining and machine learning because of the imbalance of most of the data set. An effective solution of this problem is Cost-Sensitive Learning (CSL), but when the costs are not given, this method cannot work property. As a Cost-Free Learning (CFL) method, Mutual-Information Classification (MIC) can obtain the optimal classification results when the cost information is not given. But this method emphasizes the data of minority class too much and neglects the accuracy of the classification of majority class. And based on the above, this paper presented a CFL method called Mutual-Information-ADASYN Classification (MIAC). Firstly, we get the abstaining samples which are hard to be classified by using MIC. Then we use these abstention samples to synthesize new instance by using the method of ADASYN. Thirdly, we build Mutual- Information-ADASYN Classification using the new samples. Finally, we use our classifier to get the final results. We evaluated the performance of MIAC on several imbalance binary datasets with different imbalance ratios. The experimental results indicate that the MIAC is more effective than MIC on dealing with imbalanced datasets.