Xiaohan Lin, Liu Yuan, Fangjiong Chen, Huang Yang, Xiaohu Ge
{"title":"Stochastic gradient compression for federated learning over wireless network","authors":"Xiaohan Lin, Liu Yuan, Fangjiong Chen, Huang Yang, Xiaohu Ge","doi":"10.23919/JCC.fa.2022-0660.202404","DOIUrl":null,"url":null,"abstract":"As a mature distributed machine learning paradigm, federated learning enables wireless edge devices to collaboratively train a shared AI-model by stochastic gradient descent (SGD). However, devices need to upload high-dimensional stochastic gradients to edge server in training, which cause severe communication bottleneck. To address this problem, we compress the communication by sparsifying and quantizing the stochastic gradients of edge devices. We first derive a closed form of the communication compression in terms of sparsification and quantization factors. Then, the convergence rate of this communication-compressed system is analyzed and several insights are obtained. Finally, we formulate and deal with the quantization resource allocation problem for the goal of minimizing the convergence upper bound, under the constraint of multiple-access channel capacity. Simulations show that the proposed scheme outperforms the benchmarks.","PeriodicalId":504777,"journal":{"name":"China Communications","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"China Communications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/JCC.fa.2022-0660.202404","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
As a mature distributed machine learning paradigm, federated learning enables wireless edge devices to collaboratively train a shared AI-model by stochastic gradient descent (SGD). However, devices need to upload high-dimensional stochastic gradients to edge server in training, which cause severe communication bottleneck. To address this problem, we compress the communication by sparsifying and quantizing the stochastic gradients of edge devices. We first derive a closed form of the communication compression in terms of sparsification and quantization factors. Then, the convergence rate of this communication-compressed system is analyzed and several insights are obtained. Finally, we formulate and deal with the quantization resource allocation problem for the goal of minimizing the convergence upper bound, under the constraint of multiple-access channel capacity. Simulations show that the proposed scheme outperforms the benchmarks.