Roula Nassif;Stefan Vlaski;Marco Carpentiero;Vincenzo Matta;Ali H. Sayed
{"title":"基于差分误差反馈的高效沟通分散学习","authors":"Roula Nassif;Stefan Vlaski;Marco Carpentiero;Vincenzo Matta;Ali H. Sayed","doi":"10.1109/TSP.2025.3564416","DOIUrl":null,"url":null,"abstract":"Communication-constrained algorithms for decentralized learning and optimization rely on local updates coupled with the exchange of compressed signals. In this context, <italic>differential quantization</i> is an effective technique to mitigate the negative impact of compression by leveraging correlations between successive iterates. In addition, the use of <italic>error feedback</i>, which consists of incorporating the compression error into subsequent steps, is a powerful mechanism to compensate for the bias caused by the compression. Under error feedback, performance guarantees in the literature have so far focused on algorithms employing a fusion center or a special class of contractive compressors that cannot be implemented with a finite number of bits. In this work, we propose a new <italic>decentralized</i> communication-efficient learning approach that blends differential quantization with error feedback. The approach is specifically tailored for decentralized learning problems where agents have individual risk functions to minimize subject to subspace constraints that require the minimizers across the network to lie in low-dimensional subspaces. This constrained formulation includes consensus or single-task optimization as special cases, and allows for more general task relatedness models such as multitask smoothness and coupled optimization. We show that, under some general conditions on the compression noise, and for sufficiently small step-sizes <inline-formula><tex-math>$\\mu$</tex-math></inline-formula>, the resulting communication-efficient strategy is stable both in terms of mean-square error and average bit rate: by reducing <inline-formula><tex-math>$\\mu$</tex-math></inline-formula>, it is possible to keep the <italic>estimation errors small (on the order of</i> <inline-formula><tex-math>$\\mu$</tex-math></inline-formula><italic>) without increasing indefinitely the bit rate as</i> <inline-formula><tex-math>$\\mu\\rightarrow 0$</tex-math></inline-formula>. The results establish that, in the <italic>small step-size regime</i> and with a <italic>finite number of bits</i>, it is possible to attain the performance achievable in the absence of compression.","PeriodicalId":13330,"journal":{"name":"IEEE Transactions on Signal Processing","volume":"73 ","pages":"1905-1921"},"PeriodicalIF":4.6000,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Differential Error Feedback for Communication-Efficient Decentralized Learning\",\"authors\":\"Roula Nassif;Stefan Vlaski;Marco Carpentiero;Vincenzo Matta;Ali H. Sayed\",\"doi\":\"10.1109/TSP.2025.3564416\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Communication-constrained algorithms for decentralized learning and optimization rely on local updates coupled with the exchange of compressed signals. In this context, <italic>differential quantization</i> is an effective technique to mitigate the negative impact of compression by leveraging correlations between successive iterates. In addition, the use of <italic>error feedback</i>, which consists of incorporating the compression error into subsequent steps, is a powerful mechanism to compensate for the bias caused by the compression. Under error feedback, performance guarantees in the literature have so far focused on algorithms employing a fusion center or a special class of contractive compressors that cannot be implemented with a finite number of bits. In this work, we propose a new <italic>decentralized</i> communication-efficient learning approach that blends differential quantization with error feedback. The approach is specifically tailored for decentralized learning problems where agents have individual risk functions to minimize subject to subspace constraints that require the minimizers across the network to lie in low-dimensional subspaces. This constrained formulation includes consensus or single-task optimization as special cases, and allows for more general task relatedness models such as multitask smoothness and coupled optimization. We show that, under some general conditions on the compression noise, and for sufficiently small step-sizes <inline-formula><tex-math>$\\\\mu$</tex-math></inline-formula>, the resulting communication-efficient strategy is stable both in terms of mean-square error and average bit rate: by reducing <inline-formula><tex-math>$\\\\mu$</tex-math></inline-formula>, it is possible to keep the <italic>estimation errors small (on the order of</i> <inline-formula><tex-math>$\\\\mu$</tex-math></inline-formula><italic>) without increasing indefinitely the bit rate as</i> <inline-formula><tex-math>$\\\\mu\\\\rightarrow 0$</tex-math></inline-formula>. The results establish that, in the <italic>small step-size regime</i> and with a <italic>finite number of bits</i>, it is possible to attain the performance achievable in the absence of compression.\",\"PeriodicalId\":13330,\"journal\":{\"name\":\"IEEE Transactions on Signal Processing\",\"volume\":\"73 \",\"pages\":\"1905-1921\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2025-04-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Signal Processing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10976577/\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10976577/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Differential Error Feedback for Communication-Efficient Decentralized Learning
Communication-constrained algorithms for decentralized learning and optimization rely on local updates coupled with the exchange of compressed signals. In this context, differential quantization is an effective technique to mitigate the negative impact of compression by leveraging correlations between successive iterates. In addition, the use of error feedback, which consists of incorporating the compression error into subsequent steps, is a powerful mechanism to compensate for the bias caused by the compression. Under error feedback, performance guarantees in the literature have so far focused on algorithms employing a fusion center or a special class of contractive compressors that cannot be implemented with a finite number of bits. In this work, we propose a new decentralized communication-efficient learning approach that blends differential quantization with error feedback. The approach is specifically tailored for decentralized learning problems where agents have individual risk functions to minimize subject to subspace constraints that require the minimizers across the network to lie in low-dimensional subspaces. This constrained formulation includes consensus or single-task optimization as special cases, and allows for more general task relatedness models such as multitask smoothness and coupled optimization. We show that, under some general conditions on the compression noise, and for sufficiently small step-sizes $\mu$, the resulting communication-efficient strategy is stable both in terms of mean-square error and average bit rate: by reducing $\mu$, it is possible to keep the estimation errors small (on the order of $\mu$) without increasing indefinitely the bit rate as $\mu\rightarrow 0$. The results establish that, in the small step-size regime and with a finite number of bits, it is possible to attain the performance achievable in the absence of compression.
期刊介绍:
The IEEE Transactions on Signal Processing covers novel theory, algorithms, performance analyses and applications of techniques for the processing, understanding, learning, retrieval, mining, and extraction of information from signals. The term “signal” includes, among others, audio, video, speech, image, communication, geophysical, sonar, radar, medical and musical signals. Examples of topics of interest include, but are not limited to, information processing and the theory and application of filtering, coding, transmitting, estimating, detecting, analyzing, recognizing, synthesizing, recording, and reproducing signals.