{"title":"HyperMem: Hypernetwork with memory for forgetting problem in federated reinforcement learning","authors":"Suhang Wei, Xiang Feng, Yang Xu, Huiqun Yu","doi":"10.1016/j.eswa.2025.128671","DOIUrl":null,"url":null,"abstract":"<div><div>Federated reinforcement learning plays a crucial role in decentralized and privacy-preserving policy optimization but is challenged by task heterogeneity and client dropout. Several approaches proposed for these issues, but few consider their combined impact. In this paper, we reveal the catastrophic forgetting phenomenon arising from their coexistence, which significantly degrades the global model’s performance on offline clients. We formally define this forgetting problem and establish an exponential convergence rate for hypernetwork-based federated learning methods, highlighting the adverse effects of embedding length on forgetting. Furthermore, we demonstrate the equivalence between the mean squared error loss and the chain rule in hypernetwork updates, introducing a novel updating paradigm. Based on our theoretical insights, we propose HyperMem with three key components: (1) Constrained Principal Component Embedding, which limits embedding length and enhances hypernetwork priors; (2) In-cluster and Out-cluster Losses, designed under the new updating paradigm to dynamically select fitting targets and mitigate, or even resolve, forgetting problems; and (3) Adapter Pool, enabling federated training of structurally heterogeneous client models caused by greater task heterogeneity. Comprehensive experiments demonstrate that HyperMem effectively overcomes the forgetting problem, improving training performance by 14.95 % compared to state-of-the-art methods. We implemented HyperMem as a pluggable Spark service for practical applications, reducing job runtime by 42.38 % and communication costs by 98.6 %, while ensuring data security.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"294 ","pages":"Article 128671"},"PeriodicalIF":7.5000,"publicationDate":"2025-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417425022894","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Federated reinforcement learning plays a crucial role in decentralized and privacy-preserving policy optimization but is challenged by task heterogeneity and client dropout. Several approaches proposed for these issues, but few consider their combined impact. In this paper, we reveal the catastrophic forgetting phenomenon arising from their coexistence, which significantly degrades the global model’s performance on offline clients. We formally define this forgetting problem and establish an exponential convergence rate for hypernetwork-based federated learning methods, highlighting the adverse effects of embedding length on forgetting. Furthermore, we demonstrate the equivalence between the mean squared error loss and the chain rule in hypernetwork updates, introducing a novel updating paradigm. Based on our theoretical insights, we propose HyperMem with three key components: (1) Constrained Principal Component Embedding, which limits embedding length and enhances hypernetwork priors; (2) In-cluster and Out-cluster Losses, designed under the new updating paradigm to dynamically select fitting targets and mitigate, or even resolve, forgetting problems; and (3) Adapter Pool, enabling federated training of structurally heterogeneous client models caused by greater task heterogeneity. Comprehensive experiments demonstrate that HyperMem effectively overcomes the forgetting problem, improving training performance by 14.95 % compared to state-of-the-art methods. We implemented HyperMem as a pluggable Spark service for practical applications, reducing job runtime by 42.38 % and communication costs by 98.6 %, while ensuring data security.
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.