Richeng Jin;Xiaofan He;Caijun Zhong;Zhaoyang Zhang;Tony Q. S. Quek;Huaiyu Dai
{"title":"幅度很重要:在数据异构的情况下,通过幅度感知的稀疏化和误差反馈修复 SIGNSGD","authors":"Richeng Jin;Xiaofan He;Caijun Zhong;Zhaoyang Zhang;Tony Q. S. Quek;Huaiyu Dai","doi":"10.1109/TSP.2024.3454986","DOIUrl":null,"url":null,"abstract":"Communication overhead has become one of the major bottlenecks in the distributed training of deep neural networks. To alleviate the concern, various gradient compression methods have been proposed, and sign-based algorithms are of surging interest. However, \n<sc>sign</small>\nSGD fails to converge in the presence of data heterogeneity, which is commonly observed in the emerging federated learning (FL) paradigm. Error feedback has been proposed to address the non-convergence issue. Nonetheless, it requires the workers to locally keep track of the compression errors, which renders it not suitable for FL since the workers may not participate in the training throughout the learning process. In this paper, we propose a magnitude-driven sparsification scheme, which addresses the non-convergence issue of \n<sc>sign</small>\nSGD while further improving communication efficiency. Moreover, the local update and the error feedback schemes are further incorporated to improve the learning performance (i.e., test accuracy and communication efficiency), and the convergence of the proposed method is established. The effectiveness of the proposed scheme is validated through extensive experiments on Fashion-MNIST, CIFAR-10, CIFAR-100, Tiny-ImageNet, and Mini-ImageNet datasets.","PeriodicalId":13330,"journal":{"name":"IEEE Transactions on Signal Processing","volume":"72 ","pages":"5140-5155"},"PeriodicalIF":4.6000,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Magnitude Matters: Fixing signSGD Through Magnitude-Aware Sparsification and Error Feedback in the Presence of Data Heterogeneity\",\"authors\":\"Richeng Jin;Xiaofan He;Caijun Zhong;Zhaoyang Zhang;Tony Q. S. Quek;Huaiyu Dai\",\"doi\":\"10.1109/TSP.2024.3454986\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Communication overhead has become one of the major bottlenecks in the distributed training of deep neural networks. To alleviate the concern, various gradient compression methods have been proposed, and sign-based algorithms are of surging interest. However, \\n<sc>sign</small>\\nSGD fails to converge in the presence of data heterogeneity, which is commonly observed in the emerging federated learning (FL) paradigm. Error feedback has been proposed to address the non-convergence issue. Nonetheless, it requires the workers to locally keep track of the compression errors, which renders it not suitable for FL since the workers may not participate in the training throughout the learning process. In this paper, we propose a magnitude-driven sparsification scheme, which addresses the non-convergence issue of \\n<sc>sign</small>\\nSGD while further improving communication efficiency. Moreover, the local update and the error feedback schemes are further incorporated to improve the learning performance (i.e., test accuracy and communication efficiency), and the convergence of the proposed method is established. The effectiveness of the proposed scheme is validated through extensive experiments on Fashion-MNIST, CIFAR-10, CIFAR-100, Tiny-ImageNet, and Mini-ImageNet datasets.\",\"PeriodicalId\":13330,\"journal\":{\"name\":\"IEEE Transactions on Signal Processing\",\"volume\":\"72 \",\"pages\":\"5140-5155\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2024-09-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Signal Processing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10685055/\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10685055/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Magnitude Matters: Fixing signSGD Through Magnitude-Aware Sparsification and Error Feedback in the Presence of Data Heterogeneity
Communication overhead has become one of the major bottlenecks in the distributed training of deep neural networks. To alleviate the concern, various gradient compression methods have been proposed, and sign-based algorithms are of surging interest. However,
sign
SGD fails to converge in the presence of data heterogeneity, which is commonly observed in the emerging federated learning (FL) paradigm. Error feedback has been proposed to address the non-convergence issue. Nonetheless, it requires the workers to locally keep track of the compression errors, which renders it not suitable for FL since the workers may not participate in the training throughout the learning process. In this paper, we propose a magnitude-driven sparsification scheme, which addresses the non-convergence issue of
sign
SGD while further improving communication efficiency. Moreover, the local update and the error feedback schemes are further incorporated to improve the learning performance (i.e., test accuracy and communication efficiency), and the convergence of the proposed method is established. The effectiveness of the proposed scheme is validated through extensive experiments on Fashion-MNIST, CIFAR-10, CIFAR-100, Tiny-ImageNet, and Mini-ImageNet datasets.
期刊介绍:
The IEEE Transactions on Signal Processing covers novel theory, algorithms, performance analyses and applications of techniques for the processing, understanding, learning, retrieval, mining, and extraction of information from signals. The term “signal” includes, among others, audio, video, speech, image, communication, geophysical, sonar, radar, medical and musical signals. Examples of topics of interest include, but are not limited to, information processing and the theory and application of filtering, coding, transmitting, estimating, detecting, analyzing, recognizing, synthesizing, recording, and reproducing signals.