深度学习上下文敏感可视化模型的评价

A. Dunn, D. Inkpen, Razvan Andonie
{"title":"深度学习上下文敏感可视化模型的评价","authors":"A. Dunn, D. Inkpen, Razvan Andonie","doi":"10.1109/IV56949.2022.00066","DOIUrl":null,"url":null,"abstract":"The introduction of Transformer neural networks has changed the landscape of Natural Language Processing (NLP) during the recent years. These models are very complex, and therefore hard to debug and explain. In this context, visual explanation became an attractive approach. The visualization of the path that leads to certain outputs of a model is at the core of visual explanation, as this illuminates the features or parts of the model that may need to be changed to achieve the desired results. In particular, one goal of a NLP visual explanation is to highlight the most significant parts of the text that have the greatest impact on the model output. Several visual explanation methods for NLP models were recently proposed. A major challenge is how to compare the performances of such methods since we cannot simply use the usual classification accuracy measures to evaluate the quality of visualizations. We need good metrics and rigorous criteria to measure how useful the extracted knowledge is for explaining the models. In addition, we want to visualize the differences between the knowledge extracted by different models, in order to be able to rank them. In this paper, we investigate how to evaluate explanations/visualizations resulted from machine learning models for text classification. The goal is not to improve the accuracy of a particular NLP classifier, but to assess the quality of the visualizations that explain its decisions. We describe several methods for evaluating the quality of NLP visualizations, including both automated techniques based on quantifiable measures and subjective techniques based on human judgements.","PeriodicalId":153161,"journal":{"name":"2022 26th International Conference Information Visualisation (IV)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Evaluation of Deep Learning Context-Sensitive Visualization Models\",\"authors\":\"A. Dunn, D. Inkpen, Razvan Andonie\",\"doi\":\"10.1109/IV56949.2022.00066\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The introduction of Transformer neural networks has changed the landscape of Natural Language Processing (NLP) during the recent years. These models are very complex, and therefore hard to debug and explain. In this context, visual explanation became an attractive approach. The visualization of the path that leads to certain outputs of a model is at the core of visual explanation, as this illuminates the features or parts of the model that may need to be changed to achieve the desired results. In particular, one goal of a NLP visual explanation is to highlight the most significant parts of the text that have the greatest impact on the model output. Several visual explanation methods for NLP models were recently proposed. A major challenge is how to compare the performances of such methods since we cannot simply use the usual classification accuracy measures to evaluate the quality of visualizations. We need good metrics and rigorous criteria to measure how useful the extracted knowledge is for explaining the models. In addition, we want to visualize the differences between the knowledge extracted by different models, in order to be able to rank them. In this paper, we investigate how to evaluate explanations/visualizations resulted from machine learning models for text classification. The goal is not to improve the accuracy of a particular NLP classifier, but to assess the quality of the visualizations that explain its decisions. We describe several methods for evaluating the quality of NLP visualizations, including both automated techniques based on quantifiable measures and subjective techniques based on human judgements.\",\"PeriodicalId\":153161,\"journal\":{\"name\":\"2022 26th International Conference Information Visualisation (IV)\",\"volume\":\"52 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 26th International Conference Information Visualisation (IV)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IV56949.2022.00066\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 26th International Conference Information Visualisation (IV)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IV56949.2022.00066","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

近年来,变压器神经网络的引入改变了自然语言处理(NLP)的格局。这些模型非常复杂,因此很难调试和解释。在这种情况下,视觉解释成为一种有吸引力的方法。导致模型的某些输出的路径的可视化是可视化解释的核心,因为它阐明了可能需要更改以实现预期结果的模型的特征或部分。特别是,NLP可视化解释的一个目标是突出文本中对模型输出影响最大的最重要部分。最近提出了几种NLP模型的可视化解释方法。一个主要的挑战是如何比较这些方法的性能,因为我们不能简单地使用通常的分类精度度量来评估可视化的质量。我们需要良好的度量和严格的标准来衡量提取的知识对解释模型的有用程度。此外,我们希望可视化不同模型提取的知识之间的差异,以便能够对它们进行排序。在本文中,我们研究了如何评估机器学习模型对文本分类的解释/可视化结果。目标不是提高特定NLP分类器的准确性,而是评估解释其决策的可视化的质量。我们描述了几种评估NLP可视化质量的方法,包括基于可量化测量的自动化技术和基于人类判断的主观技术。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Evaluation of Deep Learning Context-Sensitive Visualization Models
The introduction of Transformer neural networks has changed the landscape of Natural Language Processing (NLP) during the recent years. These models are very complex, and therefore hard to debug and explain. In this context, visual explanation became an attractive approach. The visualization of the path that leads to certain outputs of a model is at the core of visual explanation, as this illuminates the features or parts of the model that may need to be changed to achieve the desired results. In particular, one goal of a NLP visual explanation is to highlight the most significant parts of the text that have the greatest impact on the model output. Several visual explanation methods for NLP models were recently proposed. A major challenge is how to compare the performances of such methods since we cannot simply use the usual classification accuracy measures to evaluate the quality of visualizations. We need good metrics and rigorous criteria to measure how useful the extracted knowledge is for explaining the models. In addition, we want to visualize the differences between the knowledge extracted by different models, in order to be able to rank them. In this paper, we investigate how to evaluate explanations/visualizations resulted from machine learning models for text classification. The goal is not to improve the accuracy of a particular NLP classifier, but to assess the quality of the visualizations that explain its decisions. We describe several methods for evaluating the quality of NLP visualizations, including both automated techniques based on quantifiable measures and subjective techniques based on human judgements.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信