Optimization and analysis of distributed averaging with memory

Boris N. Oreshkin, M. Coates, M. Rabbat
{"title":"Optimization and analysis of distributed averaging with memory","authors":"Boris N. Oreshkin, M. Coates, M. Rabbat","doi":"10.1109/ALLERTON.2009.5394786","DOIUrl":null,"url":null,"abstract":"This paper analyzes the rate of convergence of a distributed averaging scheme making use of memory at each node. In conventional distributed averaging, each node computes an update based on its current state and the current states of their neighbours. Previous work observed the trajectories at each node converge smoothly and demonstrated via simulation that a predictive framework can lead to faster rates of convergence. This paper provides theoretical guarantees for a distributed averaging algorithm with memory. We analyze a scheme where updates are computed as a convex combination of two terms: (i) the usual update using only current states, and (ii) a local linear predictor term that makes use of a node's current and previous states. Although this scheme only requires one additional memory register, we prove that this approach can lead to dramatic improvements in the rate of convergence. For example, on the N-node chain topology, our approach leads to a factor of N improvement over the standard approach, and on the two-dimensional grid, our approach achieves a factor of √N improvement. Our analysis is direct and involves relating the eigenvalues of a conventional (memoryless) averaging matrix to the eigenvalues of the averaging matrix implementing the proposed scheme via a standard linearization of the quadratic eigenvalue problem. The success of our approach relies on each node using the optimal parameter for combining the two update terms. We derive a closed form expression for the optimal parameter as a function of the second largest eigenvalue of a memoryless averaging matrix, which can easily be computed in a decentralized fashion using existing methods, making our approach amenable to a practical implementation.","PeriodicalId":440015,"journal":{"name":"2009 47th Annual Allerton Conference on Communication, Control, and Computing (Allerton)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2009-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2009 47th Annual Allerton Conference on Communication, Control, and Computing (Allerton)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ALLERTON.2009.5394786","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

This paper analyzes the rate of convergence of a distributed averaging scheme making use of memory at each node. In conventional distributed averaging, each node computes an update based on its current state and the current states of their neighbours. Previous work observed the trajectories at each node converge smoothly and demonstrated via simulation that a predictive framework can lead to faster rates of convergence. This paper provides theoretical guarantees for a distributed averaging algorithm with memory. We analyze a scheme where updates are computed as a convex combination of two terms: (i) the usual update using only current states, and (ii) a local linear predictor term that makes use of a node's current and previous states. Although this scheme only requires one additional memory register, we prove that this approach can lead to dramatic improvements in the rate of convergence. For example, on the N-node chain topology, our approach leads to a factor of N improvement over the standard approach, and on the two-dimensional grid, our approach achieves a factor of √N improvement. Our analysis is direct and involves relating the eigenvalues of a conventional (memoryless) averaging matrix to the eigenvalues of the averaging matrix implementing the proposed scheme via a standard linearization of the quadratic eigenvalue problem. The success of our approach relies on each node using the optimal parameter for combining the two update terms. We derive a closed form expression for the optimal parameter as a function of the second largest eigenvalue of a memoryless averaging matrix, which can easily be computed in a decentralized fashion using existing methods, making our approach amenable to a practical implementation.
内存分布式平均算法的优化与分析
本文分析了一种利用各节点内存的分布式平均方案的收敛速度。在传统的分布式平均中,每个节点根据其当前状态及其邻居的当前状态计算更新。先前的工作观察到每个节点的轨迹平滑收敛,并通过模拟证明了预测框架可以导致更快的收敛速度。本文为具有内存的分布式平均算法提供了理论保证。我们分析了一个方案,其中更新是作为两个项的凸组合计算的:(i)仅使用当前状态的常规更新,以及(ii)利用节点当前和以前状态的局部线性预测项。虽然这种方案只需要一个额外的内存寄存器,但我们证明了这种方法可以显著提高收敛速度。例如,在N节点链拓扑上,我们的方法比标准方法提高了N倍,而在二维网格上,我们的方法提高了√N倍。我们的分析是直接的,涉及将传统(无记忆)平均矩阵的特征值与通过二次特征值问题的标准线性化实现所提出方案的平均矩阵的特征值联系起来。我们方法的成功依赖于每个节点使用最优参数来组合两个更新项。我们导出了最优参数的封闭形式表达式,作为无记忆平均矩阵的第二大特征值的函数,可以使用现有方法以分散的方式轻松计算,使我们的方法适用于实际实现。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信