VQA模型的关注和误差诱导输入区域的生成和评价解释

Arijit Ray, Michael Cogswell, Xiaoyu Lin, Kamran Alipour, Ajay Divakaran, Yi Yao, Giedrius Burachas
{"title":"VQA模型的关注和误差诱导输入区域的生成和评价解释","authors":"Arijit Ray, Michael Cogswell, Xiaoyu Lin, Kamran Alipour, Ajay Divakaran, Yi Yao, Giedrius Burachas","doi":"10.22541/au.162464902.28050142/v1","DOIUrl":null,"url":null,"abstract":"Attention maps, a popular heatmap-based explanation method for Visual\nQuestion Answering (VQA), are supposed to help users understand the\nmodel by highlighting portions of the image/question used by the model\nto infer answers. However, we see that users are often misled by current\nattention map visualizations that point to relevant regions despite the\nmodel producing an incorrect answer. Hence, we propose Error Maps that\nclarify the error by highlighting image regions where the model is prone\nto err. Error maps can indicate when a correctly attended region may be\nprocessed incorrectly leading to an incorrect answer, and hence, improve\nusers’ understanding of those cases. To evaluate our new explanations,\nwe further introduce a metric that simulates users’ interpretation of\nexplanations to evaluate their potential helpfulness to understand model\ncorrectness. We finally conduct user studies to see that our new\nexplanations help users understand model correctness better than\nbaselines by an expected 30% and that our proxy helpfulness metrics\ncorrelate strongly (rho>0.97) with how well users can\npredict model correctness.","PeriodicalId":72253,"journal":{"name":"Applied AI letters","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Generating and Evaluating Explanations of Attended and Error-Inducing Input Regions for VQA Models\",\"authors\":\"Arijit Ray, Michael Cogswell, Xiaoyu Lin, Kamran Alipour, Ajay Divakaran, Yi Yao, Giedrius Burachas\",\"doi\":\"10.22541/au.162464902.28050142/v1\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Attention maps, a popular heatmap-based explanation method for Visual\\nQuestion Answering (VQA), are supposed to help users understand the\\nmodel by highlighting portions of the image/question used by the model\\nto infer answers. However, we see that users are often misled by current\\nattention map visualizations that point to relevant regions despite the\\nmodel producing an incorrect answer. Hence, we propose Error Maps that\\nclarify the error by highlighting image regions where the model is prone\\nto err. Error maps can indicate when a correctly attended region may be\\nprocessed incorrectly leading to an incorrect answer, and hence, improve\\nusers’ understanding of those cases. To evaluate our new explanations,\\nwe further introduce a metric that simulates users’ interpretation of\\nexplanations to evaluate their potential helpfulness to understand model\\ncorrectness. We finally conduct user studies to see that our new\\nexplanations help users understand model correctness better than\\nbaselines by an expected 30% and that our proxy helpfulness metrics\\ncorrelate strongly (rho>0.97) with how well users can\\npredict model correctness.\",\"PeriodicalId\":72253,\"journal\":{\"name\":\"Applied AI letters\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-03-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied AI letters\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.22541/au.162464902.28050142/v1\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied AI letters","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.22541/au.162464902.28050142/v1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

注意力图是一种流行的基于热图的视觉问答(VQA)解释方法,旨在通过突出显示模型用于推断答案的图像/问题的部分来帮助用户理解模型。然而,我们看到,尽管模型产生了错误的答案,但用户经常被当前指向相关区域的注意力地图可视化所误导。因此,我们提出了误差图,通过突出显示模型容易出错的图像区域来澄清误差。错误图可以指示正确参与的区域何时可能被错误处理,从而导致错误答案,从而提高用户对这些情况的理解。为了评估我们的新解释,我们进一步引入了一个指标,模拟用户对解释的解释,以评估他们对理解模型正确性的潜在帮助。最后,我们对用户进行了研究,发现我们的新解释可以帮助用户比基线更好地理解模型的正确性,预期的正确率为30%,并且我们的代理有用性指标与用户预测模型正确性的程度强相关(rho>0.97)。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Generating and Evaluating Explanations of Attended and Error-Inducing Input Regions for VQA Models
Attention maps, a popular heatmap-based explanation method for Visual Question Answering (VQA), are supposed to help users understand the model by highlighting portions of the image/question used by the model to infer answers. However, we see that users are often misled by current attention map visualizations that point to relevant regions despite the model producing an incorrect answer. Hence, we propose Error Maps that clarify the error by highlighting image regions where the model is prone to err. Error maps can indicate when a correctly attended region may be processed incorrectly leading to an incorrect answer, and hence, improve users’ understanding of those cases. To evaluate our new explanations, we further introduce a metric that simulates users’ interpretation of explanations to evaluate their potential helpfulness to understand model correctness. We finally conduct user studies to see that our new explanations help users understand model correctness better than baselines by an expected 30% and that our proxy helpfulness metrics correlate strongly (rho>0.97) with how well users can predict model correctness.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信