Zhenyu Nie , Zheng Xiao , Tao Wang , Anthony Theodore Chronopoulos , Răzvan Andonie , Amir Mosavi
{"title":"提示:用于学习模型解释的文本交互评价指标","authors":"Zhenyu Nie , Zheng Xiao , Tao Wang , Anthony Theodore Chronopoulos , Răzvan Andonie , Amir Mosavi","doi":"10.1016/j.eswa.2025.128184","DOIUrl":null,"url":null,"abstract":"<div><div>Explaining the decision-making behavior of deep neural networks (DNNs) can increase their trustworthiness in real-world applications. For natural language processing (NLP) tasks, many existing interpretation methods split the text according to the interactions between words. Also, the evaluation of explanation capability focuses on justifying the importance of the divided text spans from the perspective of interaction contribution. However, the prior evaluations are misled by extra interactions, making the evaluation unable to acquire accurate interactions within the text spans. Besides, existing research considers only absolute interaction contribution, which causes the evaluation to underestimate the important text spans with lower absolute interaction contribution and to overestimate the unimportant text spans with higher absolute interaction contribution. In this work, we propose a metric called Text Interaction Proportional Score (TIPS) to evaluate faithful interpretation methods. More specifically, we use a pick scheme to acquire the interactions within the divided text span and eliminate the influence of the extra interactions. Meanwhile, we utilize the relative interaction contribution between the divided text span and whole text to measure the importance of the acquired interactions. The proposed metric is validated using two interpretation methods in explaining three neural text classifiers (LSTM, CNN and BERT) on six benchmark datasets. Experiments show that TIPS outperforms a baseline method in three ways consistently and significantly (i.e., acquiring interactions within the text span, measuring importance of interaction, and distinguishing the important and unimportant text spans).</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"287 ","pages":"Article 128184"},"PeriodicalIF":7.5000,"publicationDate":"2025-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"TIPS: A text interaction evaluation metric for learning model interpretation\",\"authors\":\"Zhenyu Nie , Zheng Xiao , Tao Wang , Anthony Theodore Chronopoulos , Răzvan Andonie , Amir Mosavi\",\"doi\":\"10.1016/j.eswa.2025.128184\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Explaining the decision-making behavior of deep neural networks (DNNs) can increase their trustworthiness in real-world applications. For natural language processing (NLP) tasks, many existing interpretation methods split the text according to the interactions between words. Also, the evaluation of explanation capability focuses on justifying the importance of the divided text spans from the perspective of interaction contribution. However, the prior evaluations are misled by extra interactions, making the evaluation unable to acquire accurate interactions within the text spans. Besides, existing research considers only absolute interaction contribution, which causes the evaluation to underestimate the important text spans with lower absolute interaction contribution and to overestimate the unimportant text spans with higher absolute interaction contribution. In this work, we propose a metric called Text Interaction Proportional Score (TIPS) to evaluate faithful interpretation methods. More specifically, we use a pick scheme to acquire the interactions within the divided text span and eliminate the influence of the extra interactions. Meanwhile, we utilize the relative interaction contribution between the divided text span and whole text to measure the importance of the acquired interactions. The proposed metric is validated using two interpretation methods in explaining three neural text classifiers (LSTM, CNN and BERT) on six benchmark datasets. Experiments show that TIPS outperforms a baseline method in three ways consistently and significantly (i.e., acquiring interactions within the text span, measuring importance of interaction, and distinguishing the important and unimportant text spans).</div></div>\",\"PeriodicalId\":50461,\"journal\":{\"name\":\"Expert Systems with Applications\",\"volume\":\"287 \",\"pages\":\"Article 128184\"},\"PeriodicalIF\":7.5000,\"publicationDate\":\"2025-05-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Expert Systems with Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0957417425018044\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417425018044","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
TIPS: A text interaction evaluation metric for learning model interpretation
Explaining the decision-making behavior of deep neural networks (DNNs) can increase their trustworthiness in real-world applications. For natural language processing (NLP) tasks, many existing interpretation methods split the text according to the interactions between words. Also, the evaluation of explanation capability focuses on justifying the importance of the divided text spans from the perspective of interaction contribution. However, the prior evaluations are misled by extra interactions, making the evaluation unable to acquire accurate interactions within the text spans. Besides, existing research considers only absolute interaction contribution, which causes the evaluation to underestimate the important text spans with lower absolute interaction contribution and to overestimate the unimportant text spans with higher absolute interaction contribution. In this work, we propose a metric called Text Interaction Proportional Score (TIPS) to evaluate faithful interpretation methods. More specifically, we use a pick scheme to acquire the interactions within the divided text span and eliminate the influence of the extra interactions. Meanwhile, we utilize the relative interaction contribution between the divided text span and whole text to measure the importance of the acquired interactions. The proposed metric is validated using two interpretation methods in explaining three neural text classifiers (LSTM, CNN and BERT) on six benchmark datasets. Experiments show that TIPS outperforms a baseline method in three ways consistently and significantly (i.e., acquiring interactions within the text span, measuring importance of interaction, and distinguishing the important and unimportant text spans).
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.