On the atout ticket learning problem for neural networks and its application in securing federated learning exchanges

IF 3.8 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS
Abdelhak Bouayad , Mohammed Akallouch , Abdelkader El Mahdaouy , Hamza Alami , Ismail Berrada
{"title":"On the atout ticket learning problem for neural networks and its application in securing federated learning exchanges","authors":"Abdelhak Bouayad ,&nbsp;Mohammed Akallouch ,&nbsp;Abdelkader El Mahdaouy ,&nbsp;Hamza Alami ,&nbsp;Ismail Berrada","doi":"10.1016/j.jisa.2024.103891","DOIUrl":null,"url":null,"abstract":"<div><div>Artificial Neural Networks (ANNs) have become the backbone of many real-world applications, including distributed applications relying on Federated Learning (FL). However, several vulnerabilities/attacks have emerged in recent years, affecting the benefits of using ANNs in FL, such as reconstruction attacks and membership inference attacks. These attacks can have severe impacts on both the societal and professional levels. For instance, inferring the presence of a patient’s private health record in a medical study or a clinic database violates the patient’s privacy and can have legal or ethical consequences. Therefore, protecting the data and model from malicious attacks in FL systems is important. This paper introduces the Atout Ticket Learning (ATL) problem. This new problem consists of identifying sensitive parameters (atout tickets) of a neural network model, which, if modified, will increase the model’s loss by at least a given threshold <span><math><mi>ϵ</mi></math></span>. First, we formulate ATL as an <span><math><msub><mrow><mi>ℓ</mi></mrow><mrow><mn>0</mn></mrow></msub></math></span>-norm minimization problem, and we derive a lower bound on the number of atout tickets needed to achieve a model degradation of <span><math><mi>ϵ</mi></math></span>. Second, we design the Atout Ticket Protocol (ATP) as an effective solution for privacy-preserving in FL systems using atout tickets, along with the benefit of noise perturbations and secure aggregation techniques. Finally, we experiment ATP against FL reconstruction attacks using new selection strategies, namely Inverting Gradients, Deep Leakage, and Improved Deep Leakage. The results show that ATP is highly robust against these attacks.</div></div>","PeriodicalId":48638,"journal":{"name":"Journal of Information Security and Applications","volume":"86 ","pages":"Article 103891"},"PeriodicalIF":3.8000,"publicationDate":"2024-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Information Security and Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2214212624001935","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Artificial Neural Networks (ANNs) have become the backbone of many real-world applications, including distributed applications relying on Federated Learning (FL). However, several vulnerabilities/attacks have emerged in recent years, affecting the benefits of using ANNs in FL, such as reconstruction attacks and membership inference attacks. These attacks can have severe impacts on both the societal and professional levels. For instance, inferring the presence of a patient’s private health record in a medical study or a clinic database violates the patient’s privacy and can have legal or ethical consequences. Therefore, protecting the data and model from malicious attacks in FL systems is important. This paper introduces the Atout Ticket Learning (ATL) problem. This new problem consists of identifying sensitive parameters (atout tickets) of a neural network model, which, if modified, will increase the model’s loss by at least a given threshold ϵ. First, we formulate ATL as an 0-norm minimization problem, and we derive a lower bound on the number of atout tickets needed to achieve a model degradation of ϵ. Second, we design the Atout Ticket Protocol (ATP) as an effective solution for privacy-preserving in FL systems using atout tickets, along with the benefit of noise perturbations and secure aggregation techniques. Finally, we experiment ATP against FL reconstruction attacks using new selection strategies, namely Inverting Gradients, Deep Leakage, and Improved Deep Leakage. The results show that ATP is highly robust against these attacks.
论神经网络的出票学习问题及其在保障联合学习交流中的应用
人工神经网络(ANN)已成为许多实际应用的支柱,包括依赖于联合学习(FL)的分布式应用。然而,近年来出现了一些漏洞/攻击,影响了在联机学习中使用人工神经网络的好处,如重构攻击和成员推理攻击。这些攻击会对社会和专业层面产生严重影响。例如,推断医学研究或诊所数据库中是否存在病人的私人健康记录会侵犯病人的隐私,并可能产生法律或道德后果。因此,在 FL 系统中保护数据和模型免受恶意攻击非常重要。本文介绍了无票学习(ATL)问题。这一新问题包括识别神经网络模型的敏感参数(out ticket),如果修改这些参数,将使模型的损失至少增加给定阈值ϵ。首先,我们将 ATL 表述为一个 ℓ0-norm 最小化问题,并推导出实现模型损失ϵ 所需的out ticket 数量下限。其次,我们设计了出票协议(ATP),作为在 FL 系统中使用出票保护隐私的有效解决方案,同时利用噪声扰动和安全聚合技术。最后,我们使用新的选择策略(即反梯度、深度泄漏和改进的深度泄漏)对 ATP 进行了实验,以对抗 FL 重构攻击。结果表明,ATP 对这些攻击具有很强的鲁棒性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Journal of Information Security and Applications
Journal of Information Security and Applications Computer Science-Computer Networks and Communications
CiteScore
10.90
自引率
5.40%
发文量
206
审稿时长
56 days
期刊介绍: Journal of Information Security and Applications (JISA) focuses on the original research and practice-driven applications with relevance to information security and applications. JISA provides a common linkage between a vibrant scientific and research community and industry professionals by offering a clear view on modern problems and challenges in information security, as well as identifying promising scientific and "best-practice" solutions. JISA issues offer a balance between original research work and innovative industrial approaches by internationally renowned information security experts and researchers.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信