Ziwei Zhan , Weijie Liu , Xiaoxi Zhang , Chee Wei Tan , Lei Xue , Haisheng Tan , Xu Chen
{"title":"通过动态梯度替代和客户选择加速个性化联邦学习","authors":"Ziwei Zhan , Weijie Liu , Xiaoxi Zhang , Chee Wei Tan , Lei Xue , Haisheng Tan , Xu Chen","doi":"10.1016/j.comnet.2025.111428","DOIUrl":null,"url":null,"abstract":"<div><div>Personalized federated learning (PFL) has gained widespread attention for its ability to preserve privacy and adapt to user-specific characteristics. Among the leading PFL methods, meta-learning based algorithms like Per-FedAvg offer a unified framework of gradient updates for all clients, eliminating the necessity of personalized model architectures that are common in other PFL approaches. However, their computation inefficiency and challenges in accommodating system heterogeneity are under-explored. This work proposes <em>pFedSara</em>, a novel PFL framework that accelerates the training of a target PFL method, Per-FedAvg, by exploiting the lightweight, vanilla FL algorithm, FedAvg. Instead of fervently creating marginally altered approaches, <em>pFedSara</em> is the first that strategically <em>reuses and blends</em> existing techniques for PFL training, navigating the runtime-accuracy trade-off, and it offers a comprehensive theoretical analysis. Specifically, it leverages dynamic gradient substitution and client selection by assessing runtime, loss, and gradient similarity between FedAvg and Per-FedAvg, the two candidate local update methods for each client. Additionally, it incorporates gradient scaling to accommodate incomplete Per-FedAvg computations that cannot be replaced by FedAvg, eliminating additional biases. A novel convergence analysis is provided, quantifying the biases introduced by both heterogeneous data and our employed hybrid update methods for computation speed-up. Extensive experiments demonstrate that <em>pFedSara</em> achieves superior training efficiency compared with state-of-the-art PFL methods.</div></div>","PeriodicalId":50637,"journal":{"name":"Computer Networks","volume":"270 ","pages":"Article 111428"},"PeriodicalIF":4.6000,"publicationDate":"2025-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Accelerating personalized federated learning via dynamic gradient substitution and client selection\",\"authors\":\"Ziwei Zhan , Weijie Liu , Xiaoxi Zhang , Chee Wei Tan , Lei Xue , Haisheng Tan , Xu Chen\",\"doi\":\"10.1016/j.comnet.2025.111428\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Personalized federated learning (PFL) has gained widespread attention for its ability to preserve privacy and adapt to user-specific characteristics. Among the leading PFL methods, meta-learning based algorithms like Per-FedAvg offer a unified framework of gradient updates for all clients, eliminating the necessity of personalized model architectures that are common in other PFL approaches. However, their computation inefficiency and challenges in accommodating system heterogeneity are under-explored. This work proposes <em>pFedSara</em>, a novel PFL framework that accelerates the training of a target PFL method, Per-FedAvg, by exploiting the lightweight, vanilla FL algorithm, FedAvg. Instead of fervently creating marginally altered approaches, <em>pFedSara</em> is the first that strategically <em>reuses and blends</em> existing techniques for PFL training, navigating the runtime-accuracy trade-off, and it offers a comprehensive theoretical analysis. Specifically, it leverages dynamic gradient substitution and client selection by assessing runtime, loss, and gradient similarity between FedAvg and Per-FedAvg, the two candidate local update methods for each client. Additionally, it incorporates gradient scaling to accommodate incomplete Per-FedAvg computations that cannot be replaced by FedAvg, eliminating additional biases. A novel convergence analysis is provided, quantifying the biases introduced by both heterogeneous data and our employed hybrid update methods for computation speed-up. Extensive experiments demonstrate that <em>pFedSara</em> achieves superior training efficiency compared with state-of-the-art PFL methods.</div></div>\",\"PeriodicalId\":50637,\"journal\":{\"name\":\"Computer Networks\",\"volume\":\"270 \",\"pages\":\"Article 111428\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2025-07-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer Networks\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1389128625003950\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1389128625003950","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
Accelerating personalized federated learning via dynamic gradient substitution and client selection
Personalized federated learning (PFL) has gained widespread attention for its ability to preserve privacy and adapt to user-specific characteristics. Among the leading PFL methods, meta-learning based algorithms like Per-FedAvg offer a unified framework of gradient updates for all clients, eliminating the necessity of personalized model architectures that are common in other PFL approaches. However, their computation inefficiency and challenges in accommodating system heterogeneity are under-explored. This work proposes pFedSara, a novel PFL framework that accelerates the training of a target PFL method, Per-FedAvg, by exploiting the lightweight, vanilla FL algorithm, FedAvg. Instead of fervently creating marginally altered approaches, pFedSara is the first that strategically reuses and blends existing techniques for PFL training, navigating the runtime-accuracy trade-off, and it offers a comprehensive theoretical analysis. Specifically, it leverages dynamic gradient substitution and client selection by assessing runtime, loss, and gradient similarity between FedAvg and Per-FedAvg, the two candidate local update methods for each client. Additionally, it incorporates gradient scaling to accommodate incomplete Per-FedAvg computations that cannot be replaced by FedAvg, eliminating additional biases. A novel convergence analysis is provided, quantifying the biases introduced by both heterogeneous data and our employed hybrid update methods for computation speed-up. Extensive experiments demonstrate that pFedSara achieves superior training efficiency compared with state-of-the-art PFL methods.
期刊介绍:
Computer Networks is an international, archival journal providing a publication vehicle for complete coverage of all topics of interest to those involved in the computer communications networking area. The audience includes researchers, managers and operators of networks as well as designers and implementors. The Editorial Board will consider any material for publication that is of interest to those groups.