The Theory of Probabilistic Hierarchical Learning for Classification

Q2 Computer Science
Ziauddin Ursani, Ahsan Ahmad Ursani
{"title":"The Theory of Probabilistic Hierarchical Learning for Classification","authors":"Ziauddin Ursani, Ahsan Ahmad Ursani","doi":"10.33166/aetic.2023.01.005","DOIUrl":null,"url":null,"abstract":"Providing the ability of classification to computers has remained at the core of the faculty of artificial intelligence. Its application has now made inroads towards nearly every walk of life, spreading over healthcare, education, defence, economics, linguistics, sociology, literature, transportation, agriculture, and industry etc. To our understanding most of the problems faced by us can be formulated as classification problems. Therefore, any novel contribution in this area has a great potential of applications in the real world. This paper proposes a novel way of learning from classification datasets i.e., hierarchical learning through set partitioning. The theory of probabilistic hierarchical learning for classification has been evolved through several works while widening its scope with each instance. The theory demonstrates that the classification of any dataset can be learnt by generating a hierarchy of learnt models each capable of classifying a disjoint subset of the training set. The basic assertion behind the theory is that an accurate classification of complex datasets can be achieved through hierarchical application of low complexity models. In this paper, the theory is redefined and revised based on four mathematical principles namely, principle of successive bifurcation, principle of two-tier discrimination, principle of class membership and the principle of selective data normalization. The algorithmic implementation of each principle is also discussed. The scope of the approach is now further widened to include ten popular real-world datasets in its test base. This approach does not only produce their accurate models but also produced above 95% accuracy on average with regard to the generalising ability, which is competitive with the contemporary literature.","PeriodicalId":36440,"journal":{"name":"Annals of Emerging Technologies in Computing","volume":"277 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Annals of Emerging Technologies in Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.33166/aetic.2023.01.005","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Computer Science","Score":null,"Total":0}
引用次数: 1

Abstract

Providing the ability of classification to computers has remained at the core of the faculty of artificial intelligence. Its application has now made inroads towards nearly every walk of life, spreading over healthcare, education, defence, economics, linguistics, sociology, literature, transportation, agriculture, and industry etc. To our understanding most of the problems faced by us can be formulated as classification problems. Therefore, any novel contribution in this area has a great potential of applications in the real world. This paper proposes a novel way of learning from classification datasets i.e., hierarchical learning through set partitioning. The theory of probabilistic hierarchical learning for classification has been evolved through several works while widening its scope with each instance. The theory demonstrates that the classification of any dataset can be learnt by generating a hierarchy of learnt models each capable of classifying a disjoint subset of the training set. The basic assertion behind the theory is that an accurate classification of complex datasets can be achieved through hierarchical application of low complexity models. In this paper, the theory is redefined and revised based on four mathematical principles namely, principle of successive bifurcation, principle of two-tier discrimination, principle of class membership and the principle of selective data normalization. The algorithmic implementation of each principle is also discussed. The scope of the approach is now further widened to include ten popular real-world datasets in its test base. This approach does not only produce their accurate models but also produced above 95% accuracy on average with regard to the generalising ability, which is competitive with the contemporary literature.
分类的概率层次学习理论
为计算机提供分类能力一直是人工智能学科的核心。它的应用现在几乎渗透到生活的各个方面,包括医疗保健、教育、国防、经济学、语言学、社会学、文学、交通、农业和工业等。根据我们的理解,我们面临的大多数问题都可以表述为分类问题。因此,在这一领域的任何新贡献在现实世界中都有很大的应用潜力。本文提出了一种新的从分类数据集学习的方法,即通过集合划分进行分层学习。分类的概率层次学习理论是经过几次工作发展起来的,它的范围随着每一个实例的扩展而不断扩大。该理论表明,任何数据集的分类都可以通过生成学习模型的层次结构来学习,每个模型都能够对训练集的不相交子集进行分类。该理论背后的基本主张是,可以通过低复杂性模型的分层应用来实现复杂数据集的准确分类。本文根据连续分岔原则、两层判别原则、类隶属性原则和选择性数据归一化原则四个数学原则对该理论进行了重新定义和修正。并讨论了各原理的算法实现。该方法的范围现在进一步扩大,在其测试库中包括十个流行的真实世界数据集。这种方法不仅产生了准确的模型,而且在泛化能力方面平均准确率达到95%以上,与当代文献具有竞争力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Annals of Emerging Technologies in Computing
Annals of Emerging Technologies in Computing Computer Science-Computer Science (all)
CiteScore
3.50
自引率
0.00%
发文量
26
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信