联邦工作负载感知量化框架,用于数据敏感应用程序中的安全学习

IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS
Manu Narula , Jasraj Meena , Dinesh Kumar Vishwakarma
{"title":"联邦工作负载感知量化框架,用于数据敏感应用程序中的安全学习","authors":"Manu Narula ,&nbsp;Jasraj Meena ,&nbsp;Dinesh Kumar Vishwakarma","doi":"10.1016/j.future.2025.107772","DOIUrl":null,"url":null,"abstract":"<div><div>Federated Learning (FL) emerged as a leading secure, distributed learning technology based on sharing insights instead of data. The privacy-ensuring capability of FL has enabled its extensive use in Data-Sensitive Applications like healthcare and finance. However, the transmitted insights are at risk of leakage as the security of the medium cannot be guaranteed and can lead to the inference of the user data. Quantization is sometimes used to change these transmitted values to provide security but at the cost of accuracy loss in global models. Coupled with client dropouts, this increases performance loss. In this paper, we propose a Federated Workload-Aware Framework with Linear Quantization (Fed-WALQ), which layers the quantization process with an active client-selection technique based on the sustainable workload of the clients. The framework minimizes the dropout rates and compensates for the loss due to quantization. Through numerical experiments compared against traditional FL and Quantization-enabled FL over multiple datasets, the Fed-WALQ shows improvements in security over the former and accuracy over the latter. The accuracy improvement varies with the complexities of the involved datasets, while a substantial drop in straggler node percentages is seen in all cases (up to 91.8% drop).</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"168 ","pages":"Article 107772"},"PeriodicalIF":6.2000,"publicationDate":"2025-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Federated workload-aware quantized framework for secure learning in data-sensitive applications\",\"authors\":\"Manu Narula ,&nbsp;Jasraj Meena ,&nbsp;Dinesh Kumar Vishwakarma\",\"doi\":\"10.1016/j.future.2025.107772\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Federated Learning (FL) emerged as a leading secure, distributed learning technology based on sharing insights instead of data. The privacy-ensuring capability of FL has enabled its extensive use in Data-Sensitive Applications like healthcare and finance. However, the transmitted insights are at risk of leakage as the security of the medium cannot be guaranteed and can lead to the inference of the user data. Quantization is sometimes used to change these transmitted values to provide security but at the cost of accuracy loss in global models. Coupled with client dropouts, this increases performance loss. In this paper, we propose a Federated Workload-Aware Framework with Linear Quantization (Fed-WALQ), which layers the quantization process with an active client-selection technique based on the sustainable workload of the clients. The framework minimizes the dropout rates and compensates for the loss due to quantization. Through numerical experiments compared against traditional FL and Quantization-enabled FL over multiple datasets, the Fed-WALQ shows improvements in security over the former and accuracy over the latter. The accuracy improvement varies with the complexities of the involved datasets, while a substantial drop in straggler node percentages is seen in all cases (up to 91.8% drop).</div></div>\",\"PeriodicalId\":55132,\"journal\":{\"name\":\"Future Generation Computer Systems-The International Journal of Escience\",\"volume\":\"168 \",\"pages\":\"Article 107772\"},\"PeriodicalIF\":6.2000,\"publicationDate\":\"2025-02-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Future Generation Computer Systems-The International Journal of Escience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167739X25000676\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X25000676","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0

摘要

联邦学习(FL)是一种领先的安全、分布式学习技术,基于共享见解而不是数据。FL的隐私保护功能使其在医疗保健和金融等数据敏感应用程序中得到广泛应用。然而,由于无法保证媒介的安全性,传输的洞察存在泄漏的风险,并可能导致用户数据的推断。量化有时用于改变这些传输值以提供安全性,但代价是全局模型中的准确性损失。再加上客户端退出,这会增加性能损失。在本文中,我们提出了一个具有线性量化的联邦工作负载感知框架(Fed-WALQ),它使用基于客户端可持续工作负载的主动客户端选择技术对量化过程进行分层。该框架最大限度地降低了失分率,并补偿了量化造成的损失。通过在多个数据集上与传统FL和量化FL进行比较的数值实验,Fed-WALQ在安全性和准确性方面都比前者有所提高。准确度的提高随所涉及数据集的复杂性而变化,而在所有情况下,离散节点百分比都大幅下降(下降幅度高达91.8%)。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Federated workload-aware quantized framework for secure learning in data-sensitive applications
Federated Learning (FL) emerged as a leading secure, distributed learning technology based on sharing insights instead of data. The privacy-ensuring capability of FL has enabled its extensive use in Data-Sensitive Applications like healthcare and finance. However, the transmitted insights are at risk of leakage as the security of the medium cannot be guaranteed and can lead to the inference of the user data. Quantization is sometimes used to change these transmitted values to provide security but at the cost of accuracy loss in global models. Coupled with client dropouts, this increases performance loss. In this paper, we propose a Federated Workload-Aware Framework with Linear Quantization (Fed-WALQ), which layers the quantization process with an active client-selection technique based on the sustainable workload of the clients. The framework minimizes the dropout rates and compensates for the loss due to quantization. Through numerical experiments compared against traditional FL and Quantization-enabled FL over multiple datasets, the Fed-WALQ shows improvements in security over the former and accuracy over the latter. The accuracy improvement varies with the complexities of the involved datasets, while a substantial drop in straggler node percentages is seen in all cases (up to 91.8% drop).
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
19.90
自引率
2.70%
发文量
376
审稿时长
10.6 months
期刊介绍: Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications. Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration. Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信