{"title":"Curiosity-Driven Reinforced Learning of Undesired Actions in Autonomous Intelligent Agents","authors":"Christopher Rosser, Khalid H. Abed","doi":"10.1109/SAMI50585.2021.9378666","DOIUrl":null,"url":null,"abstract":"Autonomous exploring agents are encouraged to explore unknown states in an environment when equipped with an intrinsic motivating factor such as curiosity. Although intrinsic motivation is a useful mechanism for an autonomous exploring agent in an environment that provides sparse rewards, it doubles as a mechanism for causing the agents to act in undesirable ways. In this paper, we show that highly-curious agents, attached with neural networks trained with the Machine Learning Agent Toolkit's (ML-Agents) implementation of the Proximal Policy Optimization (PPO) algorithm, and Intrinsic Curiosity Module (ICM), learn undesirable or reckless behaviors relatively early in the training process. We also show that strong correlations in the PPO training statistics of misbehaving agents may indicate when an actual human should intervene for safety during the RL training process.","PeriodicalId":402414,"journal":{"name":"2021 IEEE 19th World Symposium on Applied Machine Intelligence and Informatics (SAMI)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 19th World Symposium on Applied Machine Intelligence and Informatics (SAMI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SAMI50585.2021.9378666","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Autonomous exploring agents are encouraged to explore unknown states in an environment when equipped with an intrinsic motivating factor such as curiosity. Although intrinsic motivation is a useful mechanism for an autonomous exploring agent in an environment that provides sparse rewards, it doubles as a mechanism for causing the agents to act in undesirable ways. In this paper, we show that highly-curious agents, attached with neural networks trained with the Machine Learning Agent Toolkit's (ML-Agents) implementation of the Proximal Policy Optimization (PPO) algorithm, and Intrinsic Curiosity Module (ICM), learn undesirable or reckless behaviors relatively early in the training process. We also show that strong correlations in the PPO training statistics of misbehaving agents may indicate when an actual human should intervene for safety during the RL training process.