使用一阶逻辑规则的可解释神经网络分类模型

IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Haiming Tuo, Zuqiang Meng, Zihao Shi, Daosheng Zhang
{"title":"使用一阶逻辑规则的可解释神经网络分类模型","authors":"Haiming Tuo,&nbsp;Zuqiang Meng,&nbsp;Zihao Shi,&nbsp;Daosheng Zhang","doi":"10.1016/j.neucom.2024.128840","DOIUrl":null,"url":null,"abstract":"<div><div>Over the past decade, the field of neural networks has made significant strides, particularly in deep learning. However, their limited interpretability has constrained their application in certain critical domains, drawing widespread criticism. Researchers have proposed various methods for explaining neural networks to address this challenge. This paper focuses on rule-based explanations for neural network classification problems. We propose IRCnet, a scalable classification model based on first-order logic rules. IRCnet consists of layers for learning conjunction and disjunction rules, utilizing binary logic activation functions to enhance interpretability. The model is initially trained using a continuous-weight version, which is later binarized to produce a discrete-weight version. During training, we innovatively employed gradient approximation method to handle the non-differentiable weight binarization function, thereby enabling the training of split matrices used for binarization. Finally, Conjunctive Normal Form (CNF) or Disjunctive Normal Form (DNF) rules are extracted from the model’s discrete-weight version. Experimental results indicate that our model achieves the highest or near-highest performance across various classification metrics in multiple structured datasets while demonstrating significant scalability. It effectively balances classification accuracy with the complexity of the generated rules.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128840"},"PeriodicalIF":5.5000,"publicationDate":"2024-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Interpretable neural network classification model using first-order logic rules\",\"authors\":\"Haiming Tuo,&nbsp;Zuqiang Meng,&nbsp;Zihao Shi,&nbsp;Daosheng Zhang\",\"doi\":\"10.1016/j.neucom.2024.128840\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Over the past decade, the field of neural networks has made significant strides, particularly in deep learning. However, their limited interpretability has constrained their application in certain critical domains, drawing widespread criticism. Researchers have proposed various methods for explaining neural networks to address this challenge. This paper focuses on rule-based explanations for neural network classification problems. We propose IRCnet, a scalable classification model based on first-order logic rules. IRCnet consists of layers for learning conjunction and disjunction rules, utilizing binary logic activation functions to enhance interpretability. The model is initially trained using a continuous-weight version, which is later binarized to produce a discrete-weight version. During training, we innovatively employed gradient approximation method to handle the non-differentiable weight binarization function, thereby enabling the training of split matrices used for binarization. Finally, Conjunctive Normal Form (CNF) or Disjunctive Normal Form (DNF) rules are extracted from the model’s discrete-weight version. Experimental results indicate that our model achieves the highest or near-highest performance across various classification metrics in multiple structured datasets while demonstrating significant scalability. It effectively balances classification accuracy with the complexity of the generated rules.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"614 \",\"pages\":\"Article 128840\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2024-11-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231224016114\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231224016114","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

过去十年间,神经网络领域取得了长足的进步,尤其是在深度学习方面。然而,神经网络有限的可解释性限制了其在某些关键领域的应用,招致了广泛的批评。为了应对这一挑战,研究人员提出了各种解释神经网络的方法。本文重点关注神经网络分类问题的基于规则的解释。我们提出了基于一阶逻辑规则的可扩展分类模型 IRCnet。IRCnet 由用于学习连接和析取规则的层组成,利用二元逻辑激活函数来增强可解释性。该模型最初使用连续权重版本进行训练,然后将其二值化,生成离散权重版本。在训练过程中,我们创新性地采用了梯度逼近法来处理无差别权重二值化函数,从而实现了用于二值化的分割矩阵的训练。最后,从模型的离散权重版本中提取出连接正则表达式(CNF)或分离正则表达式(DNF)规则。实验结果表明,在多个结构化数据集中,我们的模型在各种分类指标上都取得了最高或接近最高的性能,同时还表现出显著的可扩展性。它有效地平衡了分类准确性和生成规则的复杂性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Interpretable neural network classification model using first-order logic rules
Over the past decade, the field of neural networks has made significant strides, particularly in deep learning. However, their limited interpretability has constrained their application in certain critical domains, drawing widespread criticism. Researchers have proposed various methods for explaining neural networks to address this challenge. This paper focuses on rule-based explanations for neural network classification problems. We propose IRCnet, a scalable classification model based on first-order logic rules. IRCnet consists of layers for learning conjunction and disjunction rules, utilizing binary logic activation functions to enhance interpretability. The model is initially trained using a continuous-weight version, which is later binarized to produce a discrete-weight version. During training, we innovatively employed gradient approximation method to handle the non-differentiable weight binarization function, thereby enabling the training of split matrices used for binarization. Finally, Conjunctive Normal Form (CNF) or Disjunctive Normal Form (DNF) rules are extracted from the model’s discrete-weight version. Experimental results indicate that our model achieves the highest or near-highest performance across various classification metrics in multiple structured datasets while demonstrating significant scalability. It effectively balances classification accuracy with the complexity of the generated rules.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信