Shaohua Cao , Huixin Wu , Xiwen Wu , Ruhui Ma , Danxin Wang , Zhu Han , Weishan Zhang
{"title":"FedDA: Resource-adaptive federated learning with dual-alignment aggregation optimization for heterogeneous edge devices","authors":"Shaohua Cao , Huixin Wu , Xiwen Wu , Ruhui Ma , Danxin Wang , Zhu Han , Weishan Zhang","doi":"10.1016/j.future.2024.107551","DOIUrl":null,"url":null,"abstract":"<div><div>Federated learning (FL) is an emerging distributed learning paradigm that allows multiple clients to collaborate on training a global model without sharing their local data. However, in practical heterogeneous edge device scenarios, FL faces the challenges of system heterogeneity and data heterogeneity, which leads to unfair participation and degraded global model performance. In this paper, we introduce FedDA, a resource-adaptive FL framework, which adapts to the client’s computing resources by assigning heterogeneous models of different sizes. To improve the performance of heterogeneous model aggregation and adjust to non-independent and identically distributed (non-i.i.d.) data, we propose a dual-alignment aggregation optimization method, i.e., parameter feature space alignment and output space alignment. Specifically, FedDA exploits the permutation symmetry property of weight space to permutate the model parameter positions via an adaptive layer-wise matching method, thereby aligning models with significant deviations in parameter feature space. FedDA mitigates the imbalance in parameter quantity between smaller and larger models through parameter expansion. Additionally, FedDA maps client labels into a uniform embedding space through output space alignment, thus reducing model parameter deviations due to non-i.i.d. data without additional client-side computing overhead. We evaluate the performance of FedDA on benchmark datasets, including FashionMNIST, CIFAR10, CIFAR100 and AGNews. Experimental results demonstrate that FedDA achieves up to 8.71% improvement in model accuracy compared to baseline methods, highlighting its effectiveness in addressing the challenges of heterogeneity.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"163 ","pages":"Article 107551"},"PeriodicalIF":6.2000,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X24005156","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
Federated learning (FL) is an emerging distributed learning paradigm that allows multiple clients to collaborate on training a global model without sharing their local data. However, in practical heterogeneous edge device scenarios, FL faces the challenges of system heterogeneity and data heterogeneity, which leads to unfair participation and degraded global model performance. In this paper, we introduce FedDA, a resource-adaptive FL framework, which adapts to the client’s computing resources by assigning heterogeneous models of different sizes. To improve the performance of heterogeneous model aggregation and adjust to non-independent and identically distributed (non-i.i.d.) data, we propose a dual-alignment aggregation optimization method, i.e., parameter feature space alignment and output space alignment. Specifically, FedDA exploits the permutation symmetry property of weight space to permutate the model parameter positions via an adaptive layer-wise matching method, thereby aligning models with significant deviations in parameter feature space. FedDA mitigates the imbalance in parameter quantity between smaller and larger models through parameter expansion. Additionally, FedDA maps client labels into a uniform embedding space through output space alignment, thus reducing model parameter deviations due to non-i.i.d. data without additional client-side computing overhead. We evaluate the performance of FedDA on benchmark datasets, including FashionMNIST, CIFAR10, CIFAR100 and AGNews. Experimental results demonstrate that FedDA achieves up to 8.71% improvement in model accuracy compared to baseline methods, highlighting its effectiveness in addressing the challenges of heterogeneity.
期刊介绍:
Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications.
Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration.
Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.