Kaiyu Wang , Bohao Qu , Menglin Zhang , Xianchang Wang , Ximing Li
{"title":"SUNRISE:在全噪声环境下,通过邻居观察的多智能体强化学习","authors":"Kaiyu Wang , Bohao Qu , Menglin Zhang , Xianchang Wang , Ximing Li","doi":"10.1016/j.eswa.2025.129781","DOIUrl":null,"url":null,"abstract":"<div><div>Multi-agent reinforcement learning (MARL) methodologies have achieved notable advancements across diverse domains. Despite these successes, the susceptibility of neural networks to perturbed data and the ubiquity of external attacks in real-world settings, such as sensor noise, pose challenges for MARL approaches. The pivotal issue revolves around the effective transfer of policies learned in idealized simulation environments to the complexities inherent in real-world scenarios. More precisely, when agents are unable to obtain any accurate observations of the external environment throughout the entire policy learning process, the MARL methods cannot learn effective policies. In addressing this issue, we propose a methodology wherein noisy observations from neighboring agents are utilized, with an agent’s own noisy observations serving as surrogate ground truth. This approach facilitates the learning of effective policies by MARL methods in environments characterized by pervasive noise. We design a denoising representation network to filter out the principal state information from environment data characterized by noise to mitigate the adverse effects of noise on the process of policy learning. Then, we integrate the denoising representation network with classic MARL methodologies to learn effective policies within environments characterized by pervasive noise. A series of exhaustive experimental results demonstrate the efficacy of our approach in attenuating the impact of external attacks on the optimization parameters of neural networks during the policy-learning process. Moreover, our methodology exhibits compatibility with classic MARL methods, allowing for the learning of effective policies.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"298 ","pages":"Article 129781"},"PeriodicalIF":7.5000,"publicationDate":"2025-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"SUNRISE: multi-agent reinforcement learning via neighbors’ observations under fully noisy environments\",\"authors\":\"Kaiyu Wang , Bohao Qu , Menglin Zhang , Xianchang Wang , Ximing Li\",\"doi\":\"10.1016/j.eswa.2025.129781\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Multi-agent reinforcement learning (MARL) methodologies have achieved notable advancements across diverse domains. Despite these successes, the susceptibility of neural networks to perturbed data and the ubiquity of external attacks in real-world settings, such as sensor noise, pose challenges for MARL approaches. The pivotal issue revolves around the effective transfer of policies learned in idealized simulation environments to the complexities inherent in real-world scenarios. More precisely, when agents are unable to obtain any accurate observations of the external environment throughout the entire policy learning process, the MARL methods cannot learn effective policies. In addressing this issue, we propose a methodology wherein noisy observations from neighboring agents are utilized, with an agent’s own noisy observations serving as surrogate ground truth. This approach facilitates the learning of effective policies by MARL methods in environments characterized by pervasive noise. We design a denoising representation network to filter out the principal state information from environment data characterized by noise to mitigate the adverse effects of noise on the process of policy learning. Then, we integrate the denoising representation network with classic MARL methodologies to learn effective policies within environments characterized by pervasive noise. A series of exhaustive experimental results demonstrate the efficacy of our approach in attenuating the impact of external attacks on the optimization parameters of neural networks during the policy-learning process. Moreover, our methodology exhibits compatibility with classic MARL methods, allowing for the learning of effective policies.</div></div>\",\"PeriodicalId\":50461,\"journal\":{\"name\":\"Expert Systems with Applications\",\"volume\":\"298 \",\"pages\":\"Article 129781\"},\"PeriodicalIF\":7.5000,\"publicationDate\":\"2025-09-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Expert Systems with Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0957417425033962\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417425033962","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
SUNRISE: multi-agent reinforcement learning via neighbors’ observations under fully noisy environments
Multi-agent reinforcement learning (MARL) methodologies have achieved notable advancements across diverse domains. Despite these successes, the susceptibility of neural networks to perturbed data and the ubiquity of external attacks in real-world settings, such as sensor noise, pose challenges for MARL approaches. The pivotal issue revolves around the effective transfer of policies learned in idealized simulation environments to the complexities inherent in real-world scenarios. More precisely, when agents are unable to obtain any accurate observations of the external environment throughout the entire policy learning process, the MARL methods cannot learn effective policies. In addressing this issue, we propose a methodology wherein noisy observations from neighboring agents are utilized, with an agent’s own noisy observations serving as surrogate ground truth. This approach facilitates the learning of effective policies by MARL methods in environments characterized by pervasive noise. We design a denoising representation network to filter out the principal state information from environment data characterized by noise to mitigate the adverse effects of noise on the process of policy learning. Then, we integrate the denoising representation network with classic MARL methodologies to learn effective policies within environments characterized by pervasive noise. A series of exhaustive experimental results demonstrate the efficacy of our approach in attenuating the impact of external attacks on the optimization parameters of neural networks during the policy-learning process. Moreover, our methodology exhibits compatibility with classic MARL methods, allowing for the learning of effective policies.
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.