{"title":"Explainable Neural Networks: Achieving Interpretability in Neural Models","authors":"Manomita Chakraborty","doi":"10.1007/s11831-024-10089-4","DOIUrl":null,"url":null,"abstract":"<div><p>Data mining is the most widely used method for discovering knowledge. There are numerous data mining tasks, with classification being the most frequently encountered task in various application domains such as fraud detection, disease diagnosis, text classification, and so on. Many classification techniques, such as Bayesian classifiers, decision trees, genetic algorithms, neural networks (NNs), and so on, are available to help researchers solve problems in a variety of domains. However, NNs are the most frequently used classification approach because they are effective at solving classification problems that cannot be divided into linear and non-linear categories, have high classification accuracy on large datasets, and require minimal processing effort. Despite having good classification performances, NNs have a pitfall associated with them which hinders their applicability in some real-world applications. NNs are black boxes in nature, which means they cannot make transparent decisions that humans can interpret. Because of this limitation, NNs are unsuitable for many applications that require transparency in decision-making as well as high accuracy, such as audit mining or medical diagnosis. The well-known solution to this inherent disadvantage of NNs is to extract explainable decision rules from them. The extracted rules provide a detailed understanding of how NNs work in a human-readable format. Rule extraction is a well-established technique with a plethora of literature on the subject. However, there are very few papers whose primary goal is to survey the existing literature. As a result, the goal of this work is to provide a detailed analysis of the existing literature and to create a framework for existing and new researchers to conduct research in this field. The paper examines the state-of art from the perspective of designing framework of the algorithms, evaluation criteria, and applications.</p></div>","PeriodicalId":55473,"journal":{"name":"Archives of Computational Methods in Engineering","volume":"31 6","pages":"3535 - 3550"},"PeriodicalIF":9.7000,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Archives of Computational Methods in Engineering","FirstCategoryId":"5","ListUrlMain":"https://link.springer.com/article/10.1007/s11831-024-10089-4","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0
Abstract
Data mining is the most widely used method for discovering knowledge. There are numerous data mining tasks, with classification being the most frequently encountered task in various application domains such as fraud detection, disease diagnosis, text classification, and so on. Many classification techniques, such as Bayesian classifiers, decision trees, genetic algorithms, neural networks (NNs), and so on, are available to help researchers solve problems in a variety of domains. However, NNs are the most frequently used classification approach because they are effective at solving classification problems that cannot be divided into linear and non-linear categories, have high classification accuracy on large datasets, and require minimal processing effort. Despite having good classification performances, NNs have a pitfall associated with them which hinders their applicability in some real-world applications. NNs are black boxes in nature, which means they cannot make transparent decisions that humans can interpret. Because of this limitation, NNs are unsuitable for many applications that require transparency in decision-making as well as high accuracy, such as audit mining or medical diagnosis. The well-known solution to this inherent disadvantage of NNs is to extract explainable decision rules from them. The extracted rules provide a detailed understanding of how NNs work in a human-readable format. Rule extraction is a well-established technique with a plethora of literature on the subject. However, there are very few papers whose primary goal is to survey the existing literature. As a result, the goal of this work is to provide a detailed analysis of the existing literature and to create a framework for existing and new researchers to conduct research in this field. The paper examines the state-of art from the perspective of designing framework of the algorithms, evaluation criteria, and applications.
数据挖掘是发现知识最广泛使用的方法。数据挖掘任务繁多,其中分类是在欺诈检测、疾病诊断、文本分类等各种应用领域中最常遇到的任务。许多分类技术,如贝叶斯分类器、决策树、遗传算法、神经网络(NN)等,都可以帮助研究人员解决各种领域的问题。然而,神经网络是最常用的分类方法,因为它能有效解决无法划分为线性和非线性类别的分类问题,在大型数据集上具有较高的分类准确性,而且只需最小的处理工作量。尽管 NN 具有良好的分类性能,但与之相关的一个隐患却阻碍了它们在某些实际应用中的适用性。自然数网络本质上是一个黑盒子,这意味着它们无法做出人类可以解读的透明决策。由于这一局限性,导航网不适合许多要求决策透明和高准确性的应用,如审计挖掘或医疗诊断。众所周知,解决网络固有缺点的方法是从网络中提取可解释的决策规则。提取的规则以人类可读的格式提供了对网络如何工作的详细了解。规则提取是一项成熟的技术,相关文献不胜枚举。然而,以调查现有文献为主要目标的论文却寥寥无几。因此,这项工作的目标是对现有文献进行详细分析,并为现有研究人员和新研究人员在这一领域开展研究创建一个框架。本文从设计算法框架、评估标准和应用的角度对最新技术进行了研究。
期刊介绍:
Archives of Computational Methods in Engineering
Aim and Scope:
Archives of Computational Methods in Engineering serves as an active forum for disseminating research and advanced practices in computational engineering, particularly focusing on mechanics and related fields. The journal emphasizes extended state-of-the-art reviews in selected areas, a unique feature of its publication.
Review Format:
Reviews published in the journal offer:
A survey of current literature
Critical exposition of topics in their full complexity
By organizing the information in this manner, readers can quickly grasp the focus, coverage, and unique features of the Archives of Computational Methods in Engineering.