Lele Fu , Yuecheng Li , Sheng Huang , Chuan Chen , Chuanfu Zhang , Zibin Zheng
{"title":"Parameter-oriented contrastive schema and multi-level knowledge distillation for heterogeneous federated learning","authors":"Lele Fu , Yuecheng Li , Sheng Huang , Chuan Chen , Chuanfu Zhang , Zibin Zheng","doi":"10.1016/j.inffus.2025.103123","DOIUrl":null,"url":null,"abstract":"<div><div>Federated learning aims to unite multiple data owners to collaboratively train a machine learning model without leaking the private data. However, the non-independent identically distributed (Non-IID) data differentiates the optimization directions of different clients, thus seriously impairing the performance of global model. Most efforts handling the data heterogeneity focus on the server or client side, adopting certain strategies to mitigate the differences of local models. These single-side solutions are limited in addressing the negative impact of heterogeneous data. In this paper, we attempt to overcome the problem of heterogenous federated learning simultaneously from dual sides. Specifically, to prevent the catastrophical forgetting of global information, we devise a parameter-oriented contrastive schema for correcting the optimization directions of local models on the client-side. Furthermore, considering that the only average of very diverse network parameters might damage the structural information, a multi-level knowledge distillation manner to repair the corrupt information of the global model is performed on the server-side. A multitude of experiments on four benchmark datasets demonstrate that the proposed method outperforms the state-of-the-art federated learning approaches on the Non-IID data.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"121 ","pages":"Article 103123"},"PeriodicalIF":14.7000,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525001964","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Federated learning aims to unite multiple data owners to collaboratively train a machine learning model without leaking the private data. However, the non-independent identically distributed (Non-IID) data differentiates the optimization directions of different clients, thus seriously impairing the performance of global model. Most efforts handling the data heterogeneity focus on the server or client side, adopting certain strategies to mitigate the differences of local models. These single-side solutions are limited in addressing the negative impact of heterogeneous data. In this paper, we attempt to overcome the problem of heterogenous federated learning simultaneously from dual sides. Specifically, to prevent the catastrophical forgetting of global information, we devise a parameter-oriented contrastive schema for correcting the optimization directions of local models on the client-side. Furthermore, considering that the only average of very diverse network parameters might damage the structural information, a multi-level knowledge distillation manner to repair the corrupt information of the global model is performed on the server-side. A multitude of experiments on four benchmark datasets demonstrate that the proposed method outperforms the state-of-the-art federated learning approaches on the Non-IID data.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.