{"title":"SCDFL:一种基于谱聚类的分散联邦学习加速收敛框架","authors":"Faisal Alshami , Lin Yao , Xin Wang , Guowei Wu","doi":"10.1016/j.comnet.2025.111615","DOIUrl":null,"url":null,"abstract":"<div><div>Decentralized Federated Learning (DFL) is a popular distributed machine learning framework that facilitates collaboration among multiple clients without dependence on a central server to develop a global model. This architecture faces issues with client convergence, resulting in network congestion and slower convergence during the DFL process. These challenges stem from various communications topologies and the non-independent and non-identically distributed nature of data on terminal devices in real-world scenarios, which affect both model convergence speed and overall terminal performance. Therefore, we propose SCDFL, a federated learning framework that leverages spectral clustering to efficiently and scalably handle client data heterogeneity. SCDFL introduces a novel spectral clustering strategy that focuses on grouping clients based on their characteristics. Key components include reducing the dimensionality of the client data by incremental PCA, which includes high-dimensional model updates or feature vectors, making the clustering process more efficient. Then, a similarity matrix based on the reduced data will be computed to measure client similarity. Utilizing this matrix, we apply spectral clustering to group clients with similar data characteristics. Finally, we apply the aggregation in intra-cluster and inter-cluster to the updated global model. Extensive experiments have been conducted across different topologies, and the results demonstrate that SCDFL achieves higher accuracy, faster convergence, reduced communication overhead, and improved generalization, particularly on complex datasets like MNIST, CIFAR10, and CIFAR100, while efficiently handling data heterogeneity and optimizing resource utilization across various network topologies.</div></div>","PeriodicalId":50637,"journal":{"name":"Computer Networks","volume":"271 ","pages":"Article 111615"},"PeriodicalIF":4.6000,"publicationDate":"2025-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"SCDFL: A Spectral Clustering-based framework for accelerating convergence in Decentralized Federated Learning\",\"authors\":\"Faisal Alshami , Lin Yao , Xin Wang , Guowei Wu\",\"doi\":\"10.1016/j.comnet.2025.111615\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Decentralized Federated Learning (DFL) is a popular distributed machine learning framework that facilitates collaboration among multiple clients without dependence on a central server to develop a global model. This architecture faces issues with client convergence, resulting in network congestion and slower convergence during the DFL process. These challenges stem from various communications topologies and the non-independent and non-identically distributed nature of data on terminal devices in real-world scenarios, which affect both model convergence speed and overall terminal performance. Therefore, we propose SCDFL, a federated learning framework that leverages spectral clustering to efficiently and scalably handle client data heterogeneity. SCDFL introduces a novel spectral clustering strategy that focuses on grouping clients based on their characteristics. Key components include reducing the dimensionality of the client data by incremental PCA, which includes high-dimensional model updates or feature vectors, making the clustering process more efficient. Then, a similarity matrix based on the reduced data will be computed to measure client similarity. Utilizing this matrix, we apply spectral clustering to group clients with similar data characteristics. Finally, we apply the aggregation in intra-cluster and inter-cluster to the updated global model. Extensive experiments have been conducted across different topologies, and the results demonstrate that SCDFL achieves higher accuracy, faster convergence, reduced communication overhead, and improved generalization, particularly on complex datasets like MNIST, CIFAR10, and CIFAR100, while efficiently handling data heterogeneity and optimizing resource utilization across various network topologies.</div></div>\",\"PeriodicalId\":50637,\"journal\":{\"name\":\"Computer Networks\",\"volume\":\"271 \",\"pages\":\"Article 111615\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2025-08-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer Networks\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1389128625005821\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1389128625005821","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
SCDFL: A Spectral Clustering-based framework for accelerating convergence in Decentralized Federated Learning
Decentralized Federated Learning (DFL) is a popular distributed machine learning framework that facilitates collaboration among multiple clients without dependence on a central server to develop a global model. This architecture faces issues with client convergence, resulting in network congestion and slower convergence during the DFL process. These challenges stem from various communications topologies and the non-independent and non-identically distributed nature of data on terminal devices in real-world scenarios, which affect both model convergence speed and overall terminal performance. Therefore, we propose SCDFL, a federated learning framework that leverages spectral clustering to efficiently and scalably handle client data heterogeneity. SCDFL introduces a novel spectral clustering strategy that focuses on grouping clients based on their characteristics. Key components include reducing the dimensionality of the client data by incremental PCA, which includes high-dimensional model updates or feature vectors, making the clustering process more efficient. Then, a similarity matrix based on the reduced data will be computed to measure client similarity. Utilizing this matrix, we apply spectral clustering to group clients with similar data characteristics. Finally, we apply the aggregation in intra-cluster and inter-cluster to the updated global model. Extensive experiments have been conducted across different topologies, and the results demonstrate that SCDFL achieves higher accuracy, faster convergence, reduced communication overhead, and improved generalization, particularly on complex datasets like MNIST, CIFAR10, and CIFAR100, while efficiently handling data heterogeneity and optimizing resource utilization across various network topologies.
期刊介绍:
Computer Networks is an international, archival journal providing a publication vehicle for complete coverage of all topics of interest to those involved in the computer communications networking area. The audience includes researchers, managers and operators of networks as well as designers and implementors. The Editorial Board will consider any material for publication that is of interest to those groups.