{"title":"A Stochastic Asynchronous Gradient Descent Algorithm with Delay Compensation Mechanism","authors":"Tianyu Zhang, Tianhan Gao, Qingwei Mi","doi":"10.1109/IAICT55358.2022.9887513","DOIUrl":null,"url":null,"abstract":"A large amount of idle computing power exists in mobile devices, which can be deployed with large-scale machine learning applications. One of the key problems is how to reduce the communication overhead between different nodes. In recent years, gradient sparsity is introduced to reduce the communication overhead. However, in the federated learning scenario, the traditional synchronous gradient optimization algorithm can not adapt to the complex network environment and high communication costs. In this paper, we propose a stochastic gradient descent algorithm with delay compensation mechanism (FedDgd) for asynchronous distributed training and further optimize it for federated asynchronous training. It is proved theoretically that FedDgd can converge at the same rate as ASGD for non-convex neural networks. Moreover, FedDgd way converge quickly and tolerates staleness in various app applications as well.","PeriodicalId":154027,"journal":{"name":"2022 IEEE International Conference on Industry 4.0, Artificial Intelligence, and Communications Technology (IAICT)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Industry 4.0, Artificial Intelligence, and Communications Technology (IAICT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IAICT55358.2022.9887513","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
A large amount of idle computing power exists in mobile devices, which can be deployed with large-scale machine learning applications. One of the key problems is how to reduce the communication overhead between different nodes. In recent years, gradient sparsity is introduced to reduce the communication overhead. However, in the federated learning scenario, the traditional synchronous gradient optimization algorithm can not adapt to the complex network environment and high communication costs. In this paper, we propose a stochastic gradient descent algorithm with delay compensation mechanism (FedDgd) for asynchronous distributed training and further optimize it for federated asynchronous training. It is proved theoretically that FedDgd can converge at the same rate as ASGD for non-convex neural networks. Moreover, FedDgd way converge quickly and tolerates staleness in various app applications as well.