K. V. Ramana, Yuvasri. B, Sultanuddin Sj, P. Ponsudha, Sowmya Pd, A. V. Sangeetha
{"title":"Applying Cost-Sensitive Learning Methods to Improve Extremely Unbalanced Big Data Problems Using Random Forest","authors":"K. V. Ramana, Yuvasri. B, Sultanuddin Sj, P. Ponsudha, Sowmya Pd, A. V. Sangeetha","doi":"10.1109/ACCAI58221.2023.10199250","DOIUrl":null,"url":null,"abstract":"In a larger part minority characterization issue, class irregularity in the dataset(s) can definitely misshape the exhibition of classifiers, creating an expectation predisposition for the greater part class. A negative (larger part) class expectation predisposition could make impeding impacts if the positive (minority) class is the gathering of interest and the application region being referred to states that a false negative is significantly more costly than a false certain. The decrease of class divergence is made more troublesome by big data because of the different and muddled design of the similarly bigger datasets. This exploration presents a wide evaluation of distributed works inside the past 8 years, zeroed in on fashionable unevenness (i.e., a greater part to-minority class proportion somewhere in the range of 100:1 and 10,000:1) in big data to survey the cutting edge in addressing ominous outcomes connected with class irregularity. In this paper we propose two methods for managing the imbalanced data grouping issue utilizing irregular backwoods. The other depends on an inspecting approach, though the first depends on cost-sensitive learning. Execution pointers like review and exactness, false-positive and false-negative rates.","PeriodicalId":382104,"journal":{"name":"2023 International Conference on Advances in Computing, Communication and Applied Informatics (ACCAI)","volume":"77 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Advances in Computing, Communication and Applied Informatics (ACCAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACCAI58221.2023.10199250","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In a larger part minority characterization issue, class irregularity in the dataset(s) can definitely misshape the exhibition of classifiers, creating an expectation predisposition for the greater part class. A negative (larger part) class expectation predisposition could make impeding impacts if the positive (minority) class is the gathering of interest and the application region being referred to states that a false negative is significantly more costly than a false certain. The decrease of class divergence is made more troublesome by big data because of the different and muddled design of the similarly bigger datasets. This exploration presents a wide evaluation of distributed works inside the past 8 years, zeroed in on fashionable unevenness (i.e., a greater part to-minority class proportion somewhere in the range of 100:1 and 10,000:1) in big data to survey the cutting edge in addressing ominous outcomes connected with class irregularity. In this paper we propose two methods for managing the imbalanced data grouping issue utilizing irregular backwoods. The other depends on an inspecting approach, though the first depends on cost-sensitive learning. Execution pointers like review and exactness, false-positive and false-negative rates.