Gradient-tracking Based Differentially Private Distributed Optimization with Enhanced Optimization Accuracy

Yuanzhe Xuan, Yongqiang Wang
{"title":"Gradient-tracking Based Differentially Private Distributed Optimization with Enhanced Optimization Accuracy","authors":"Yuanzhe Xuan, Yongqiang Wang","doi":"10.48550/arXiv.2212.05364","DOIUrl":null,"url":null,"abstract":"Privacy protection has become an increasingly pressing requirement in distributed optimization. However, equipping distributed optimization with differential privacy, the state-of-the-art privacy protection mechanism, will unavoidably compromise optimization accuracy. In this paper, we propose an algorithm to achieve rigorous $\\epsilon$-differential privacy in gradient-tracking based distributed optimization with enhanced optimization accuracy. More specifically, to suppress the influence of differential-privacy noise, we propose a new robust gradient-tracking based distributed optimization algorithm that allows both stepsize and the variance of injected noise to vary with time. Then, we establish a new analyzing approach that can characterize the convergence of the gradient-tracking based algorithm under both constant and time-varying stespsizes. To our knowledge, this is the first analyzing framework that can treat gradient-tracking based distributed optimization under both constant and time-varying stepsizes in a unified manner. More importantly, the new analyzing approach gives a much less conservative analytical bound on the stepsize compared with existing proof techniques for gradient-tracking based distributed optimization. We also theoretically characterize the influence of differential-privacy design on the accuracy of distributed optimization, which reveals that inter-agent interaction has a significant impact on the final optimization accuracy. Numerical simulation results confirm the theoretical predictions.","PeriodicalId":13196,"journal":{"name":"IEEE Robotics Autom. Mag.","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Robotics Autom. Mag.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2212.05364","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Privacy protection has become an increasingly pressing requirement in distributed optimization. However, equipping distributed optimization with differential privacy, the state-of-the-art privacy protection mechanism, will unavoidably compromise optimization accuracy. In this paper, we propose an algorithm to achieve rigorous $\epsilon$-differential privacy in gradient-tracking based distributed optimization with enhanced optimization accuracy. More specifically, to suppress the influence of differential-privacy noise, we propose a new robust gradient-tracking based distributed optimization algorithm that allows both stepsize and the variance of injected noise to vary with time. Then, we establish a new analyzing approach that can characterize the convergence of the gradient-tracking based algorithm under both constant and time-varying stespsizes. To our knowledge, this is the first analyzing framework that can treat gradient-tracking based distributed optimization under both constant and time-varying stepsizes in a unified manner. More importantly, the new analyzing approach gives a much less conservative analytical bound on the stepsize compared with existing proof techniques for gradient-tracking based distributed optimization. We also theoretically characterize the influence of differential-privacy design on the accuracy of distributed optimization, which reveals that inter-agent interaction has a significant impact on the final optimization accuracy. Numerical simulation results confirm the theoretical predictions.
基于梯度跟踪的提高优化精度的差分私有分布优化
隐私保护已成为分布式优化中日益迫切的要求。然而,为分布式优化配置差分隐私这一最先进的隐私保护机制,将不可避免地影响优化的准确性。在本文中,我们提出了一种在基于梯度跟踪的分布式优化中实现严格的$\epsilon$差分隐私的算法,并提高了优化精度。更具体地说,为了抑制差分隐私噪声的影响,我们提出了一种新的基于梯度跟踪的鲁棒分布式优化算法,该算法允许步长和注入噪声的方差随时间变化。然后,我们建立了一种新的分析方法,可以表征基于梯度跟踪的算法在恒定和时变应力大小下的收敛性。据我们所知,这是第一个能够以统一的方式处理恒定和时变步长下基于梯度跟踪的分布式优化的分析框架。更重要的是,与现有的基于梯度跟踪的分布式优化证明技术相比,新的分析方法对步长给出了更小的保守分析界。我们还从理论上刻画了差分隐私设计对分布式优化精度的影响,揭示了智能体间交互对最终优化精度有显著影响。数值模拟结果证实了理论预测。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信