Global-Mapping-Consistency-Constrained Visual-Semantic Embedding for Interpreting Autonomous Perception Models

IF 4.6 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Chi Zhang;Meng Yuan;Xiaoning Ma;Ping Wei;Yuanqi Su;Li Li;Yuehu Liu
{"title":"Global-Mapping-Consistency-Constrained Visual-Semantic Embedding for Interpreting Autonomous Perception Models","authors":"Chi Zhang;Meng Yuan;Xiaoning Ma;Ping Wei;Yuanqi Su;Li Li;Yuehu Liu","doi":"10.1109/OJITS.2024.3418552","DOIUrl":null,"url":null,"abstract":"From the perspective of artificial intelligence evaluation, the need to discover and explain the potential shortness of the evaluated intelligent algorithms/systems as well as the need to evaluate the intelligence level of such testees are of equal importance. In this paper, we propose a possible solution to these challenges: Explainable Evaluation for visual intelligence. Specifically, we focus on the problem setting where the internal mechanisms of AI algorithms are sophisticated, heterogeneous or unreachable. In this case, a latent attribute dictionary learning method with constrained by mapping consistency is proposed to explain the performance variation patterns of visual perception intelligence under different test samples. By jointly iteratively solving the learning of latent concept representation for test samples and the regression of latent concept-generalization performance, the mapping relationship between deep representation, semantic attribute annotation, and generalization performance of test samples is established to predict the degree of influence of semantic attributes on visual perception generalization performance. The optimal solution of proposed method could be reached via an alternating optimization process. Through quantitative experiments, we find that global mapping consistency constraints can make the learned latent concept representation strictly consistent with deep representation, thereby improving the accuracy of semantic attribute-perception performance correlation calculation.","PeriodicalId":100631,"journal":{"name":"IEEE Open Journal of Intelligent Transportation Systems","volume":"5 ","pages":"393-408"},"PeriodicalIF":4.6000,"publicationDate":"2024-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10570287","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Open Journal of Intelligent Transportation Systems","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10570287/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

From the perspective of artificial intelligence evaluation, the need to discover and explain the potential shortness of the evaluated intelligent algorithms/systems as well as the need to evaluate the intelligence level of such testees are of equal importance. In this paper, we propose a possible solution to these challenges: Explainable Evaluation for visual intelligence. Specifically, we focus on the problem setting where the internal mechanisms of AI algorithms are sophisticated, heterogeneous or unreachable. In this case, a latent attribute dictionary learning method with constrained by mapping consistency is proposed to explain the performance variation patterns of visual perception intelligence under different test samples. By jointly iteratively solving the learning of latent concept representation for test samples and the regression of latent concept-generalization performance, the mapping relationship between deep representation, semantic attribute annotation, and generalization performance of test samples is established to predict the degree of influence of semantic attributes on visual perception generalization performance. The optimal solution of proposed method could be reached via an alternating optimization process. Through quantitative experiments, we find that global mapping consistency constraints can make the learned latent concept representation strictly consistent with deep representation, thereby improving the accuracy of semantic attribute-perception performance correlation calculation.
用于解释自主感知模型的全局映射-一致性约束视觉-语义嵌入技术
从人工智能评估的角度来看,发现和解释被评估的智能算法/系统的潜在不足以及评估这些被测试者的智能水平是同等重要的。在本文中,我们针对这些挑战提出了一种可能的解决方案:可解释的视觉智能评估。具体来说,我们将重点放在人工智能算法内部机制复杂、异构或不可触及的问题设置上。在这种情况下,我们提出了一种以映射一致性为约束的潜在属性字典学习方法,来解释视觉感知智能在不同测试样本下的性能变化规律。通过联合迭代求解测试样本的潜在概念表征学习和潜在概念-泛化性能回归,建立深度表征、语义属性标注和测试样本泛化性能之间的映射关系,预测语义属性对视知觉泛化性能的影响程度。通过交替优化过程,可以得到所提方法的最优解。通过定量实验,我们发现全局映射一致性约束可以使学习到的潜在概念表征与深层表征严格一致,从而提高语义属性与感知性能相关性计算的准确性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
5.40
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信