Constructing decision rules from naive bayes model for robust and low complexity classification

Nabeel Al-A'araji, S. Al-Mamory, Ali Al-shakarchi
{"title":"Constructing decision rules from naive bayes model for robust and low complexity classification","authors":"Nabeel Al-A'araji, S. Al-Mamory, Ali Al-shakarchi","doi":"10.26555/IJAIN.V7I1.578","DOIUrl":null,"url":null,"abstract":"Article history Selected paper from The 2020 Global Research Conference (GRaCe'20), Trengganu-Malaysia (Virtually), 16-18 October 2020, https://terengganu.uitm.edu.my/ grace2020/. Peer-reviewed by GRaCe'20 Scientific Committee and Editorial Team of IJAIN journal. Received October 26, 2020 Revised November 10, 2020 Accepted March 15, 2021 Available online March 31, 2021 A large spectrum of classifiers has been described in the literature. One attractive classification technique is a Naïve Bayes (NB) which has been relayed on probability theory. NB has two major limitations: First, it requires to rescan the dataset and applying a set of equations each time to classify instances, which is an expensive step if a dataset is relatively large. Second, NB may remain challenging for non-statisticians to understand the deep work of a model. On the other hand, Rule-Based classifiers (RBCs) have used IF-THEN rules (henceforth, rule-set), which are more comprehensible and less complex for classification tasks. For elevating NB limitations, this paper presents a method for constructing a rule-set from the NB model, which serves as RBC. Experiments of the constructing ruleset have been conducted on (Iris, WBC, Vote) datasets. Coverage, Accuracy, M-Estimate, and Laplace are crucial evaluation metrics that have been projected to rule-set. In some datasets, the rule-set obtains significant accuracy results that reach 95.33 %, 95.17% for Iris and vote datasets, respectively. The constructed rule-set can mimic the classification capability of NB, provide a visual representation of the model, express rules infidelity with acceptable accuracy; an easier method to interpreting and adjusting from the original model. Hence, the rule-set will provide a comprehensible and lightweight model than NB itself.","PeriodicalId":52195,"journal":{"name":"International Journal of Advances in Intelligent Informatics","volume":"15 12 1","pages":"76"},"PeriodicalIF":0.0000,"publicationDate":"2021-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Advances in Intelligent Informatics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.26555/IJAIN.V7I1.578","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Article history Selected paper from The 2020 Global Research Conference (GRaCe'20), Trengganu-Malaysia (Virtually), 16-18 October 2020, https://terengganu.uitm.edu.my/ grace2020/. Peer-reviewed by GRaCe'20 Scientific Committee and Editorial Team of IJAIN journal. Received October 26, 2020 Revised November 10, 2020 Accepted March 15, 2021 Available online March 31, 2021 A large spectrum of classifiers has been described in the literature. One attractive classification technique is a Naïve Bayes (NB) which has been relayed on probability theory. NB has two major limitations: First, it requires to rescan the dataset and applying a set of equations each time to classify instances, which is an expensive step if a dataset is relatively large. Second, NB may remain challenging for non-statisticians to understand the deep work of a model. On the other hand, Rule-Based classifiers (RBCs) have used IF-THEN rules (henceforth, rule-set), which are more comprehensible and less complex for classification tasks. For elevating NB limitations, this paper presents a method for constructing a rule-set from the NB model, which serves as RBC. Experiments of the constructing ruleset have been conducted on (Iris, WBC, Vote) datasets. Coverage, Accuracy, M-Estimate, and Laplace are crucial evaluation metrics that have been projected to rule-set. In some datasets, the rule-set obtains significant accuracy results that reach 95.33 %, 95.17% for Iris and vote datasets, respectively. The constructed rule-set can mimic the classification capability of NB, provide a visual representation of the model, express rules infidelity with acceptable accuracy; an easier method to interpreting and adjusting from the original model. Hence, the rule-set will provide a comprehensible and lightweight model than NB itself.
利用朴素贝叶斯模型构造决策规则,实现鲁棒性和低复杂度分类
论文选自2020年全球研究会议(GRaCe'20),马来西亚丁加努(虚拟),2020年10月16-18日,https://terengganu.uitm.edu.my/ grace2020/。由GRaCe'20科学委员会和IJAIN期刊编辑团队同行评审。收到2020年10月26日修订2020年11月10日接受2021年3月15日在线发布2021年3月31日文献中描述了大量的分类器。一种有吸引力的分类技术是Naïve贝叶斯(NB),它是在概率论的基础上发展起来的。NB有两个主要的限制:首先,它需要重新扫描数据集并每次应用一组方程来对实例进行分类,如果数据集相对较大,这是一个昂贵的步骤。其次,NB对于非统计学家来说,理解模型的深层工作可能仍然具有挑战性。另一方面,基于规则的分类器(rbc)使用IF-THEN规则(以后称为规则集),这些规则对于分类任务来说更容易理解,也不那么复杂。为了提高NB的局限性,本文提出了一种从NB模型构造规则集的方法,该规则集作为RBC。在Iris、WBC、Vote等数据集上进行了规则集构建实验。Coverage, Accuracy, M-Estimate,和Laplace是预测到规则集的关键评估指标。在一些数据集中,该规则集获得了显著的准确率结果,Iris和vote数据集的准确率分别达到95.33%、95.17%。所构建的规则集可以模拟NB的分类能力,提供模型的可视化表示,以可接受的精度表达规则的不一致性;一种更容易解释和调整原始模型的方法。因此,规则集将提供比NB本身更易于理解和轻量级的模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
International Journal of Advances in Intelligent Informatics
International Journal of Advances in Intelligent Informatics Computer Science-Computer Vision and Pattern Recognition
CiteScore
3.00
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信