Combining AI control systems and human decision support via robustness and criticality

Walt Woods, Alexander Grushin, Simon Khan, Alvaro Velasquez
{"title":"Combining AI control systems and human decision support via robustness and criticality","authors":"Walt Woods, Alexander Grushin, Simon Khan, Alvaro Velasquez","doi":"10.1117/12.3016311","DOIUrl":null,"url":null,"abstract":"AI-enabled capabilities are reaching the requisite level of maturity to be deployed in the real world. Yet, the ability of these systems to always make correct or safe decisions is a constant source of criticism and reluctance to use them. One way of addressing these concerns is to leverage AI control systems alongside and in support of human decisions, relying on the AI control system in safe situations while calling on a human co-decider for critical situations. Additionally, by leveraging an AI control system built specifically to assist in joint human/machine decisions, the opportunity naturally arises to then use human interactions to continuously improve the AI control system’s accuracy and robustness. We extend a methodology for Adversarial Explanations (AE) to state-of-the-art reinforcement learning frameworks, including MuZero. Multiple improvements to the base agent architecture are proposed. We demonstrate how this technology has two applications: for intelligent decision tools and to enhance training / learning frameworks. In a decision support context, adversarial explanations help a user make the correct decision by highlighting those contextual factors that would need to change for a different AI-recommended decision. As another benefit of adversarial explanations, we show that the learned AI control system demonstrates robustness against adversarial tampering. Additionally, we supplement AE by introducing Strategically Similar Autoencoders (SSAs) to help users identify and understand all salient factors being considered by the AI system. In a training / learning framework, this technology can improve both the AI’s decisions and explanations through human interaction. Finally, to identify when AI decisions would most benefit from human oversight, we tie this combined system to our prior art on statistically verified analyses of the criticality of decisions at any point in time.","PeriodicalId":178341,"journal":{"name":"Defense + Commercial Sensing","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Defense + Commercial Sensing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.3016311","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

AI-enabled capabilities are reaching the requisite level of maturity to be deployed in the real world. Yet, the ability of these systems to always make correct or safe decisions is a constant source of criticism and reluctance to use them. One way of addressing these concerns is to leverage AI control systems alongside and in support of human decisions, relying on the AI control system in safe situations while calling on a human co-decider for critical situations. Additionally, by leveraging an AI control system built specifically to assist in joint human/machine decisions, the opportunity naturally arises to then use human interactions to continuously improve the AI control system’s accuracy and robustness. We extend a methodology for Adversarial Explanations (AE) to state-of-the-art reinforcement learning frameworks, including MuZero. Multiple improvements to the base agent architecture are proposed. We demonstrate how this technology has two applications: for intelligent decision tools and to enhance training / learning frameworks. In a decision support context, adversarial explanations help a user make the correct decision by highlighting those contextual factors that would need to change for a different AI-recommended decision. As another benefit of adversarial explanations, we show that the learned AI control system demonstrates robustness against adversarial tampering. Additionally, we supplement AE by introducing Strategically Similar Autoencoders (SSAs) to help users identify and understand all salient factors being considered by the AI system. In a training / learning framework, this technology can improve both the AI’s decisions and explanations through human interaction. Finally, to identify when AI decisions would most benefit from human oversight, we tie this combined system to our prior art on statistically verified analyses of the criticality of decisions at any point in time.
通过鲁棒性和临界性将人工智能控制系统与人类决策支持相结合
人工智能功能正在达到在现实世界中部署所需的成熟度。然而,这些系统能否始终做出正确或安全的决策,一直是人们批评和不愿使用它们的原因。解决这些问题的一种方法是利用人工智能控制系统来辅助人类决策,在安全情况下依靠人工智能控制系统,而在危急情况下则由人类共同决策。此外,通过利用专为协助人类/机器联合决策而构建的人工智能控制系统,自然就有机会利用人类互动来不断提高人工智能控制系统的准确性和鲁棒性。我们将对抗性解释(AE)方法扩展到最先进的强化学习框架,包括 MuZero。我们对基础代理架构提出了多项改进建议。我们展示了这项技术的两种应用:智能决策工具和增强训练/学习框架。在决策支持环境中,对抗性解释通过强调那些需要改变的环境因素,帮助用户做出正确的决策,从而做出不同的人工智能推荐决策。作为对抗性解释的另一个优势,我们展示了所学人工智能控制系统在对抗性篡改方面的鲁棒性。此外,我们还通过引入策略相似自动编码器(SSA)来补充人工智能,帮助用户识别和理解人工智能系统正在考虑的所有突出因素。在训练/学习框架中,这项技术可以通过人机交互改进人工智能的决策和解释。最后,为了确定人工智能决策何时最能受益于人类的监督,我们将这一组合系统与我们在任何时间点对决策关键性进行统计验证分析的现有技术结合起来。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信