{"title":"A Chunked Local Aggregation Strategy in Federated Learning","authors":"Haibing Zhao, Weiqin Tong, Xiaoli Zhi, Tong Liu","doi":"10.1109/ICTAI56018.2022.00014","DOIUrl":null,"url":null,"abstract":"Federated Learning (FL) is a distributed machine learning technology that trains models on large-scale distributed devices while keeping training data localized and privatized. However, in settings where data is distributed in a not independent and identically distributed (non-I.I.D.) fashion, the single joint model produced by FL suffers in terms of test set accuracy and communication costs. And a multi-layer topology are widely deployed for FL in real scenarios. Therefore, we propose FedBox, a chunked local aggregation federated learning framework to improve the generalization ability and aggregation efficiency of model in non-I.I.D. data by adapting to the topology of the real network. Moreover, we study the adaptive gradient descent (AGC) to mitigate the feature shift caused by training non-I.I.D. data. In this work, we modified the aggregation strategy of FL by introducing a virtual node layer based on local stochastic gradient methods (SGD), and separate the edge node cluster by the similarity between the local update model and the global update model. We show that FedBox can effectively improve convergence speed and test accuracy, while reducing communication cost. Training results on FederatedEMNIST, Cifar10, Cifar100 and Shakespeare datasets indicate that FedBox allows model training to converge in fewer communication rounds and improves training accuracy by up to 3.1% compared with FedAVG. In addition, we make an empirical analysis of the extended range of virtual nodes.","PeriodicalId":354314,"journal":{"name":"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTAI56018.2022.00014","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Federated Learning (FL) is a distributed machine learning technology that trains models on large-scale distributed devices while keeping training data localized and privatized. However, in settings where data is distributed in a not independent and identically distributed (non-I.I.D.) fashion, the single joint model produced by FL suffers in terms of test set accuracy and communication costs. And a multi-layer topology are widely deployed for FL in real scenarios. Therefore, we propose FedBox, a chunked local aggregation federated learning framework to improve the generalization ability and aggregation efficiency of model in non-I.I.D. data by adapting to the topology of the real network. Moreover, we study the adaptive gradient descent (AGC) to mitigate the feature shift caused by training non-I.I.D. data. In this work, we modified the aggregation strategy of FL by introducing a virtual node layer based on local stochastic gradient methods (SGD), and separate the edge node cluster by the similarity between the local update model and the global update model. We show that FedBox can effectively improve convergence speed and test accuracy, while reducing communication cost. Training results on FederatedEMNIST, Cifar10, Cifar100 and Shakespeare datasets indicate that FedBox allows model training to converge in fewer communication rounds and improves training accuracy by up to 3.1% compared with FedAVG. In addition, we make an empirical analysis of the extended range of virtual nodes.