{"title":"Privacy-preserving and Byzantine-robust federated broad learning with chain-loop structure","authors":"Nan Li, Chang-E Ren, Siyao Cheng","doi":"10.1016/j.neucom.2025.129975","DOIUrl":null,"url":null,"abstract":"<div><div>Federated learning (FL) can collaboratively train a model by aggregating local models instead of aggregating raw data, which can protect privacy by ensuring that data remains on the client. However, the traditional FL still faces some challenges such as privacy leakage and the presence of Byzantine clients. We propose a privacy-preserving and Byzantine-robust federated broad learning framework with chain-loop structure i.e., PBFBL-CL, and this algorithm can simultaneously achieve protection of clients’ privacy and robustness against Byzantine attacks. In this paper, we apply Byzantine step-by-step co-validation algorithm to address the existence of Byzantine clients. We pass the aggregated model through the chain, so each client’s privacy is well protected. Moreover, PBFBL-CL can reduce the communication overhead between clients and server. Finally, we evaluate the PBFBL-CL algorithm in MNIST, Fashion-MNIST and NORB datasets, and the results show that our algorithm is better than existing FL algorithms in terms of model accuracy and training speed. Experimental results demonstrate that under the extreme scenario where Byzantine client proportion reaches 90%, the model achieves an accuracy of 89.53%, only 4.17% lower than the 93.7% accuracy observed in the ideal scenario without Byzantine clients.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"636 ","pages":"Article 129975"},"PeriodicalIF":5.5000,"publicationDate":"2025-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225006472","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Federated learning (FL) can collaboratively train a model by aggregating local models instead of aggregating raw data, which can protect privacy by ensuring that data remains on the client. However, the traditional FL still faces some challenges such as privacy leakage and the presence of Byzantine clients. We propose a privacy-preserving and Byzantine-robust federated broad learning framework with chain-loop structure i.e., PBFBL-CL, and this algorithm can simultaneously achieve protection of clients’ privacy and robustness against Byzantine attacks. In this paper, we apply Byzantine step-by-step co-validation algorithm to address the existence of Byzantine clients. We pass the aggregated model through the chain, so each client’s privacy is well protected. Moreover, PBFBL-CL can reduce the communication overhead between clients and server. Finally, we evaluate the PBFBL-CL algorithm in MNIST, Fashion-MNIST and NORB datasets, and the results show that our algorithm is better than existing FL algorithms in terms of model accuracy and training speed. Experimental results demonstrate that under the extreme scenario where Byzantine client proportion reaches 90%, the model achieves an accuracy of 89.53%, only 4.17% lower than the 93.7% accuracy observed in the ideal scenario without Byzantine clients.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.