Hongtao Li , Xinyu Li , Ximeng Liu , Bo Wang , Jie Wang , Youliang Tian
{"title":"FedSam: Enhancing federated learning accuracy with differential privacy and data heterogeneity mitigation","authors":"Hongtao Li , Xinyu Li , Ximeng Liu , Bo Wang , Jie Wang , Youliang Tian","doi":"10.1016/j.csi.2025.104019","DOIUrl":null,"url":null,"abstract":"<div><div>A large-scale model is typically trained on an extensive dataset to update its parameters and enhance its classification capabilities. However, directly using such data can raise significant privacy concerns, especially in the medical field, where datasets often contain sensitive patient information. Federated Learning (FL) offers a solution by enabling multiple parties to collaboratively train a high-performance model without sharing their raw data. Despite this, during the federated training process, attackers can still potentially extract private information from local models. To bolster privacy protections, Differential Privacy (DP) has been introduced to FL, providing stringent safeguards. However, the combination of DP and data heterogeneity can often lead to reduced model accuracy. To tackle these challenges, we introduce a sampling-memory mechanism, FedSam, which improves the accuracy of the global model while maintaining the required noise levels for differential privacy. This mechanism also mitigates the adverse effects of data heterogeneity in heterogeneous federated environments, thereby improving the global model’s overall performance. Experimental evaluations on datasets demonstrate the superiority of our approach. FedSam achieves a classification accuracy of 95.03%, significantly outperforming traditional DP-FedAvg (91.74%) under the same privacy constraints, highlighting FedSam’s robustness and efficiency.</div></div>","PeriodicalId":50635,"journal":{"name":"Computer Standards & Interfaces","volume":"95 ","pages":"Article 104019"},"PeriodicalIF":3.1000,"publicationDate":"2025-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Standards & Interfaces","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0920548925000480","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
引用次数: 0
Abstract
A large-scale model is typically trained on an extensive dataset to update its parameters and enhance its classification capabilities. However, directly using such data can raise significant privacy concerns, especially in the medical field, where datasets often contain sensitive patient information. Federated Learning (FL) offers a solution by enabling multiple parties to collaboratively train a high-performance model without sharing their raw data. Despite this, during the federated training process, attackers can still potentially extract private information from local models. To bolster privacy protections, Differential Privacy (DP) has been introduced to FL, providing stringent safeguards. However, the combination of DP and data heterogeneity can often lead to reduced model accuracy. To tackle these challenges, we introduce a sampling-memory mechanism, FedSam, which improves the accuracy of the global model while maintaining the required noise levels for differential privacy. This mechanism also mitigates the adverse effects of data heterogeneity in heterogeneous federated environments, thereby improving the global model’s overall performance. Experimental evaluations on datasets demonstrate the superiority of our approach. FedSam achieves a classification accuracy of 95.03%, significantly outperforming traditional DP-FedAvg (91.74%) under the same privacy constraints, highlighting FedSam’s robustness and efficiency.
期刊介绍:
The quality of software, well-defined interfaces (hardware and software), the process of digitalisation, and accepted standards in these fields are essential for building and exploiting complex computing, communication, multimedia and measuring systems. Standards can simplify the design and construction of individual hardware and software components and help to ensure satisfactory interworking.
Computer Standards & Interfaces is an international journal dealing specifically with these topics.
The journal
• Provides information about activities and progress on the definition of computer standards, software quality, interfaces and methods, at national, European and international levels
• Publishes critical comments on standards and standards activities
• Disseminates user''s experiences and case studies in the application and exploitation of established or emerging standards, interfaces and methods
• Offers a forum for discussion on actual projects, standards, interfaces and methods by recognised experts
• Stimulates relevant research by providing a specialised refereed medium.