Argumentative explanations for pattern-based text classifiers

IF 1.7 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Piyawat Lertvittayakumjorn, Francesca Toni
{"title":"Argumentative explanations for pattern-based text classifiers","authors":"Piyawat Lertvittayakumjorn, Francesca Toni","doi":"10.3233/aac-220004","DOIUrl":null,"url":null,"abstract":"Recent works in Explainable AI mostly address the transparency issue of black-box models or create explanations for any kind of models (i.e., they are model-agnostic), while leaving explanations of interpretable models largely underexplored. In this paper, we fill this gap by focusing on explanations for a specific interpretable model, namely pattern-based logistic regression (PLR) for binary text classification. We do so because, albeit interpretable, PLR is challenging when it comes to explanations. In particular, we found that a standard way to extract explanations from this model does not consider relations among the features, making the explanations hardly plausible to humans. Hence, we propose AXPLR, a novel explanation method using (forms of) computational argumentation to generate explanations (for outputs computed by PLR) which unearth model agreements and disagreements among the features. Specifically, we use computational argumentation as follows: we see features (patterns) in PLR as arguments in a form of quantified bipolar argumentation frameworks (QBAFs) and extract attacks and supports between arguments based on specificity of the arguments; we understand logistic regression as a gradual semantics for these QBAFs, used to determine the arguments’ dialectic strength; and we study standard properties of gradual semantics for QBAFs in the context of our argumentative re-interpretation of PLR, sanctioning its suitability for explanatory purposes. We then show how to extract intuitive explanations (for outputs computed by PLR) from the constructed QBAFs. Finally, we conduct an empirical evaluation and two experiments in the context of human-AI collaboration to demonstrate the advantages of our resulting AXPLR method.","PeriodicalId":44268,"journal":{"name":"Argument & Computation","volume":null,"pages":null},"PeriodicalIF":1.7000,"publicationDate":"2023-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Argument & Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3233/aac-220004","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Recent works in Explainable AI mostly address the transparency issue of black-box models or create explanations for any kind of models (i.e., they are model-agnostic), while leaving explanations of interpretable models largely underexplored. In this paper, we fill this gap by focusing on explanations for a specific interpretable model, namely pattern-based logistic regression (PLR) for binary text classification. We do so because, albeit interpretable, PLR is challenging when it comes to explanations. In particular, we found that a standard way to extract explanations from this model does not consider relations among the features, making the explanations hardly plausible to humans. Hence, we propose AXPLR, a novel explanation method using (forms of) computational argumentation to generate explanations (for outputs computed by PLR) which unearth model agreements and disagreements among the features. Specifically, we use computational argumentation as follows: we see features (patterns) in PLR as arguments in a form of quantified bipolar argumentation frameworks (QBAFs) and extract attacks and supports between arguments based on specificity of the arguments; we understand logistic regression as a gradual semantics for these QBAFs, used to determine the arguments’ dialectic strength; and we study standard properties of gradual semantics for QBAFs in the context of our argumentative re-interpretation of PLR, sanctioning its suitability for explanatory purposes. We then show how to extract intuitive explanations (for outputs computed by PLR) from the constructed QBAFs. Finally, we conduct an empirical evaluation and two experiments in the context of human-AI collaboration to demonstrate the advantages of our resulting AXPLR method.
基于模式的文本分类器的论证性解释
最近在可解释的人工智能方面的工作主要是解决黑箱模型的透明度问题,或者为任何类型的模型创建解释(即,它们是模型不可知论的),而对可解释模型的解释在很大程度上没有得到充分的探索。在本文中,我们通过专注于解释一个特定的可解释模型来填补这一空白,即二元文本分类的基于模式的逻辑回归(PLR)。我们这样做的原因是,尽管PLR是可解释的,但它在解释时是具有挑战性的。特别是,我们发现从该模型中提取解释的标准方法没有考虑特征之间的关系,使得解释对人类来说很难可信。因此,我们提出了AXPLR,这是一种新的解释方法,使用(形式)计算论证来生成解释(对于由PLR计算的输出),揭示特征之间的模型一致和不一致。具体来说,我们使用计算论证如下:我们将PLR中的特征(模式)视为量化双极论证框架(qbaf)形式的论证,并根据论证的特殊性提取论证之间的攻击和支持;我们将逻辑回归理解为这些qbaf的渐进语义,用于确定论点的辩证法强度;在我们对PLR的论证性重新解释的背景下,我们研究了qbaf的渐进语义的标准性质,并批准了其解释目的的适用性。然后,我们展示了如何从构建的qbaf中提取直观的解释(对于由PLR计算的输出)。最后,我们在人类-人工智能协作的背景下进行了实证评估和两个实验,以证明我们得到的AXPLR方法的优势。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Argument & Computation
Argument & Computation COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE-
CiteScore
4.10
自引率
7.10%
发文量
8
文献相关原料
公司名称 产品信息 采购帮参考价格
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信