{"title":"针对非iid数据的个性化联邦学习中的自定义模型差异","authors":"Fengrui Hao , Taihang Zhi , Tianlong Gu , Xuguang Bao","doi":"10.1016/j.knosys.2025.114522","DOIUrl":null,"url":null,"abstract":"<div><div>Federated learning (FL) is a traditional framework comprising a central server and multiple local clients. In FL, a shared global model is trained for resource-constrained computing devices while preserving data privacy. However, in certain practical applications, the shared global model may exhibit poor inference performance in local clients owing to nonindependent and nonidentically distributed (non-IID) characteristics of data. To address this issue, researchers have proposed personalized FL (PFL), which involves learning a customized model for each client to mitigate the impact of weight divergences when the training datasets are non-IID. Unfortunately, existing studies fail to reveal the inherent connection between model discrepancies and non-IID data. Herein, we focus on demonstrating the relationship between weight divergences among customized models and non-IID data, and we provide a proposition to reveal the root cause of such divergences. Additionally, based on our theoretical analysis, we introduce two novel personalized FL methods, namely, PFL with neighbor clients (PFedNC) and PFL with neighbor layers (PFedNL), to address the issue of non-IID data scenarios. Theoretical convergence analysis and extensive experiments indicate that our proposed methods outperform state-of-the-art personalized algorithms in non-IID scenarios. Specifically, PFedNC achieves up to 4 % improvement in customized model accuracy, while PFedNL yields 8 %–10 % gains over multiple baselines.</div></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":"330 ","pages":"Article 114522"},"PeriodicalIF":7.6000,"publicationDate":"2025-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Toward customized model discrepancies in personalized federated learning on non-IID data\",\"authors\":\"Fengrui Hao , Taihang Zhi , Tianlong Gu , Xuguang Bao\",\"doi\":\"10.1016/j.knosys.2025.114522\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Federated learning (FL) is a traditional framework comprising a central server and multiple local clients. In FL, a shared global model is trained for resource-constrained computing devices while preserving data privacy. However, in certain practical applications, the shared global model may exhibit poor inference performance in local clients owing to nonindependent and nonidentically distributed (non-IID) characteristics of data. To address this issue, researchers have proposed personalized FL (PFL), which involves learning a customized model for each client to mitigate the impact of weight divergences when the training datasets are non-IID. Unfortunately, existing studies fail to reveal the inherent connection between model discrepancies and non-IID data. Herein, we focus on demonstrating the relationship between weight divergences among customized models and non-IID data, and we provide a proposition to reveal the root cause of such divergences. Additionally, based on our theoretical analysis, we introduce two novel personalized FL methods, namely, PFL with neighbor clients (PFedNC) and PFL with neighbor layers (PFedNL), to address the issue of non-IID data scenarios. Theoretical convergence analysis and extensive experiments indicate that our proposed methods outperform state-of-the-art personalized algorithms in non-IID scenarios. Specifically, PFedNC achieves up to 4 % improvement in customized model accuracy, while PFedNL yields 8 %–10 % gains over multiple baselines.</div></div>\",\"PeriodicalId\":49939,\"journal\":{\"name\":\"Knowledge-Based Systems\",\"volume\":\"330 \",\"pages\":\"Article 114522\"},\"PeriodicalIF\":7.6000,\"publicationDate\":\"2025-09-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Knowledge-Based Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0950705125015618\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705125015618","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Toward customized model discrepancies in personalized federated learning on non-IID data
Federated learning (FL) is a traditional framework comprising a central server and multiple local clients. In FL, a shared global model is trained for resource-constrained computing devices while preserving data privacy. However, in certain practical applications, the shared global model may exhibit poor inference performance in local clients owing to nonindependent and nonidentically distributed (non-IID) characteristics of data. To address this issue, researchers have proposed personalized FL (PFL), which involves learning a customized model for each client to mitigate the impact of weight divergences when the training datasets are non-IID. Unfortunately, existing studies fail to reveal the inherent connection between model discrepancies and non-IID data. Herein, we focus on demonstrating the relationship between weight divergences among customized models and non-IID data, and we provide a proposition to reveal the root cause of such divergences. Additionally, based on our theoretical analysis, we introduce two novel personalized FL methods, namely, PFL with neighbor clients (PFedNC) and PFL with neighbor layers (PFedNL), to address the issue of non-IID data scenarios. Theoretical convergence analysis and extensive experiments indicate that our proposed methods outperform state-of-the-art personalized algorithms in non-IID scenarios. Specifically, PFedNC achieves up to 4 % improvement in customized model accuracy, while PFedNL yields 8 %–10 % gains over multiple baselines.
期刊介绍:
Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.