Richeng Jin;Xiaofan He;Caijun Zhong;Zhaoyang Zhang;Tony Q. S. Quek;Huaiyu Dai
{"title":"Magnitude Matters: Fixing signSGD Through Magnitude-Aware Sparsification and Error Feedback in the Presence of Data Heterogeneity","authors":"Richeng Jin;Xiaofan He;Caijun Zhong;Zhaoyang Zhang;Tony Q. S. Quek;Huaiyu Dai","doi":"10.1109/TSP.2024.3454986","DOIUrl":null,"url":null,"abstract":"Communication overhead has become one of the major bottlenecks in the distributed training of deep neural networks. To alleviate the concern, various gradient compression methods have been proposed, and sign-based algorithms are of surging interest. However, \n<sc>sign</small>\nSGD fails to converge in the presence of data heterogeneity, which is commonly observed in the emerging federated learning (FL) paradigm. Error feedback has been proposed to address the non-convergence issue. Nonetheless, it requires the workers to locally keep track of the compression errors, which renders it not suitable for FL since the workers may not participate in the training throughout the learning process. In this paper, we propose a magnitude-driven sparsification scheme, which addresses the non-convergence issue of \n<sc>sign</small>\nSGD while further improving communication efficiency. Moreover, the local update and the error feedback schemes are further incorporated to improve the learning performance (i.e., test accuracy and communication efficiency), and the convergence of the proposed method is established. The effectiveness of the proposed scheme is validated through extensive experiments on Fashion-MNIST, CIFAR-10, CIFAR-100, Tiny-ImageNet, and Mini-ImageNet datasets.","PeriodicalId":13330,"journal":{"name":"IEEE Transactions on Signal Processing","volume":"72 ","pages":"5140-5155"},"PeriodicalIF":4.6000,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10685055/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Communication overhead has become one of the major bottlenecks in the distributed training of deep neural networks. To alleviate the concern, various gradient compression methods have been proposed, and sign-based algorithms are of surging interest. However,
sign
SGD fails to converge in the presence of data heterogeneity, which is commonly observed in the emerging federated learning (FL) paradigm. Error feedback has been proposed to address the non-convergence issue. Nonetheless, it requires the workers to locally keep track of the compression errors, which renders it not suitable for FL since the workers may not participate in the training throughout the learning process. In this paper, we propose a magnitude-driven sparsification scheme, which addresses the non-convergence issue of
sign
SGD while further improving communication efficiency. Moreover, the local update and the error feedback schemes are further incorporated to improve the learning performance (i.e., test accuracy and communication efficiency), and the convergence of the proposed method is established. The effectiveness of the proposed scheme is validated through extensive experiments on Fashion-MNIST, CIFAR-10, CIFAR-100, Tiny-ImageNet, and Mini-ImageNet datasets.
期刊介绍:
The IEEE Transactions on Signal Processing covers novel theory, algorithms, performance analyses and applications of techniques for the processing, understanding, learning, retrieval, mining, and extraction of information from signals. The term “signal” includes, among others, audio, video, speech, image, communication, geophysical, sonar, radar, medical and musical signals. Examples of topics of interest include, but are not limited to, information processing and the theory and application of filtering, coding, transmitting, estimating, detecting, analyzing, recognizing, synthesizing, recording, and reproducing signals.