{"title":"鲁棒梯度差分压缩在联邦学习中的应用","authors":"Yueyao Chen, Beilun Wang, Tianyi Ma, Cheng Chen","doi":"10.1109/CSCWD57460.2023.10152826","DOIUrl":null,"url":null,"abstract":"Nowadays, federated learning has been a prevailing paradigm for large-scale distributed machine learning, which is faced with the problem of communication bottleneck. To solve this problem, recent works usually apply different compression techniques such as sparsification and quantization compressors. However, such approaches are all lossy compression and have two drawbacks. First, they could lead to information loss of the global parameter. Second, compressed parameters carrying less information would be more likely to be attacked by malicious workers than full parameters, leading to a Byzantine failure of the model. In this paper, to avoid information loss, mitigate the communication bottleneck, and at the same time tolerate popular Byzantine attacks, we propose FedGraD, which leverages gradient difference compression and combines robust aggregation rules in federated learning settings. Our experimental results on three different datasets a9a, w8a and mushrooms show good performance of our method.","PeriodicalId":51008,"journal":{"name":"Computer Supported Cooperative Work-The Journal of Collaborative Computing","volume":"3 1","pages":"1748-1753"},"PeriodicalIF":2.0000,"publicationDate":"2023-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Applying Robust Gradient Difference Compression to Federated Learning\",\"authors\":\"Yueyao Chen, Beilun Wang, Tianyi Ma, Cheng Chen\",\"doi\":\"10.1109/CSCWD57460.2023.10152826\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Nowadays, federated learning has been a prevailing paradigm for large-scale distributed machine learning, which is faced with the problem of communication bottleneck. To solve this problem, recent works usually apply different compression techniques such as sparsification and quantization compressors. However, such approaches are all lossy compression and have two drawbacks. First, they could lead to information loss of the global parameter. Second, compressed parameters carrying less information would be more likely to be attacked by malicious workers than full parameters, leading to a Byzantine failure of the model. In this paper, to avoid information loss, mitigate the communication bottleneck, and at the same time tolerate popular Byzantine attacks, we propose FedGraD, which leverages gradient difference compression and combines robust aggregation rules in federated learning settings. Our experimental results on three different datasets a9a, w8a and mushrooms show good performance of our method.\",\"PeriodicalId\":51008,\"journal\":{\"name\":\"Computer Supported Cooperative Work-The Journal of Collaborative Computing\",\"volume\":\"3 1\",\"pages\":\"1748-1753\"},\"PeriodicalIF\":2.0000,\"publicationDate\":\"2023-05-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer Supported Cooperative Work-The Journal of Collaborative Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.1109/CSCWD57460.2023.10152826\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Supported Cooperative Work-The Journal of Collaborative Computing","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1109/CSCWD57460.2023.10152826","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
Applying Robust Gradient Difference Compression to Federated Learning
Nowadays, federated learning has been a prevailing paradigm for large-scale distributed machine learning, which is faced with the problem of communication bottleneck. To solve this problem, recent works usually apply different compression techniques such as sparsification and quantization compressors. However, such approaches are all lossy compression and have two drawbacks. First, they could lead to information loss of the global parameter. Second, compressed parameters carrying less information would be more likely to be attacked by malicious workers than full parameters, leading to a Byzantine failure of the model. In this paper, to avoid information loss, mitigate the communication bottleneck, and at the same time tolerate popular Byzantine attacks, we propose FedGraD, which leverages gradient difference compression and combines robust aggregation rules in federated learning settings. Our experimental results on three different datasets a9a, w8a and mushrooms show good performance of our method.
期刊介绍:
Computer Supported Cooperative Work (CSCW): The Journal of Collaborative Computing and Work Practices is devoted to innovative research in computer-supported cooperative work (CSCW). It provides an interdisciplinary and international forum for the debate and exchange of ideas concerning theoretical, practical, technical, and social issues in CSCW.
The CSCW Journal arose in response to the growing interest in the design, implementation and use of technical systems (including computing, information, and communications technologies) which support people working cooperatively, and its scope remains to encompass the multifarious aspects of research within CSCW and related areas.
The CSCW Journal focuses on research oriented towards the development of collaborative computing technologies on the basis of studies of actual cooperative work practices (where ‘work’ is used in the wider sense). That is, it welcomes in particular submissions that (a) report on findings from ethnographic or similar kinds of in-depth fieldwork of work practices with a view to their technological implications, (b) report on empirical evaluations of the use of extant or novel technical solutions under real-world conditions, and/or (c) develop technical or conceptual frameworks for practice-oriented computing research based on previous fieldwork and evaluations.