{"title":"集成概率聚类和深度强化学习用于边缘网络中联邦学习的偏差缓解和设备异质性","authors":"Neha Singh , Mainak Adhikari","doi":"10.1016/j.jnca.2025.104259","DOIUrl":null,"url":null,"abstract":"<div><div>Federated Learning (FL) enables decentralized and collaborative training on resource-constrained Edge Devices (EDs) while preserving data privacy by avoiding raw data transmission. However, traditional FL approaches face challenges such as non-independent and identically distributed (non-IID) data, biased model aggregation due to device heterogeneity, and inefficiencies caused by stragglers during model updates. We propose a novel Hierarchical Deep Reinforcement Learning-based Probabilistic Federated Learning (Hier-FedDRL) strategy to address these limitations. This framework combines local and central Deep Reinforcement Learning (DRL) agents with a probabilistic clustering approach to manage heterogeneous devices and optimize resource allocation dynamically. Local DRL agents optimize intra-cluster operations, including training and resource distribution, while the central DRL agent oversees global model updates and inter-cluster coordination.</div><div>To ensure balanced aggregation and mitigate biases, the proposed framework employs Gaussian Mixture Models (GMMs) for clustering EDs based on their data distributions and resource characteristics. Additionally, a dynamic contribution-based aggregation technique is introduced to fairly weigh updates from diverse EDs, reducing biases in the global model. The performance of Hier-FedDRL is evaluated in a cloud-based setup, where Docker containers are used to simulate EDs and Google Kubernetes Engine clusters for cloud orchestration. Experimental results over benchmark datasets demonstrate that the proposed Hier-FedDRL achieves 4%–6% higher accuracy, reduces convergence time by 7%–10%, and lowers bias in the global model by 25%, outperforming state-of-the-art FL approaches while effectively addressing data and resource heterogeneity.</div></div>","PeriodicalId":54784,"journal":{"name":"Journal of Network and Computer Applications","volume":"242 ","pages":"Article 104259"},"PeriodicalIF":7.7000,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Integrated probabilistic clustering and Deep Reinforcement Learning for bias mitigation and device heterogeneity of Federated Learning in edge networks\",\"authors\":\"Neha Singh , Mainak Adhikari\",\"doi\":\"10.1016/j.jnca.2025.104259\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Federated Learning (FL) enables decentralized and collaborative training on resource-constrained Edge Devices (EDs) while preserving data privacy by avoiding raw data transmission. However, traditional FL approaches face challenges such as non-independent and identically distributed (non-IID) data, biased model aggregation due to device heterogeneity, and inefficiencies caused by stragglers during model updates. We propose a novel Hierarchical Deep Reinforcement Learning-based Probabilistic Federated Learning (Hier-FedDRL) strategy to address these limitations. This framework combines local and central Deep Reinforcement Learning (DRL) agents with a probabilistic clustering approach to manage heterogeneous devices and optimize resource allocation dynamically. Local DRL agents optimize intra-cluster operations, including training and resource distribution, while the central DRL agent oversees global model updates and inter-cluster coordination.</div><div>To ensure balanced aggregation and mitigate biases, the proposed framework employs Gaussian Mixture Models (GMMs) for clustering EDs based on their data distributions and resource characteristics. Additionally, a dynamic contribution-based aggregation technique is introduced to fairly weigh updates from diverse EDs, reducing biases in the global model. The performance of Hier-FedDRL is evaluated in a cloud-based setup, where Docker containers are used to simulate EDs and Google Kubernetes Engine clusters for cloud orchestration. Experimental results over benchmark datasets demonstrate that the proposed Hier-FedDRL achieves 4%–6% higher accuracy, reduces convergence time by 7%–10%, and lowers bias in the global model by 25%, outperforming state-of-the-art FL approaches while effectively addressing data and resource heterogeneity.</div></div>\",\"PeriodicalId\":54784,\"journal\":{\"name\":\"Journal of Network and Computer Applications\",\"volume\":\"242 \",\"pages\":\"Article 104259\"},\"PeriodicalIF\":7.7000,\"publicationDate\":\"2025-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Network and Computer Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1084804525001560\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Network and Computer Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1084804525001560","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
Integrated probabilistic clustering and Deep Reinforcement Learning for bias mitigation and device heterogeneity of Federated Learning in edge networks
Federated Learning (FL) enables decentralized and collaborative training on resource-constrained Edge Devices (EDs) while preserving data privacy by avoiding raw data transmission. However, traditional FL approaches face challenges such as non-independent and identically distributed (non-IID) data, biased model aggregation due to device heterogeneity, and inefficiencies caused by stragglers during model updates. We propose a novel Hierarchical Deep Reinforcement Learning-based Probabilistic Federated Learning (Hier-FedDRL) strategy to address these limitations. This framework combines local and central Deep Reinforcement Learning (DRL) agents with a probabilistic clustering approach to manage heterogeneous devices and optimize resource allocation dynamically. Local DRL agents optimize intra-cluster operations, including training and resource distribution, while the central DRL agent oversees global model updates and inter-cluster coordination.
To ensure balanced aggregation and mitigate biases, the proposed framework employs Gaussian Mixture Models (GMMs) for clustering EDs based on their data distributions and resource characteristics. Additionally, a dynamic contribution-based aggregation technique is introduced to fairly weigh updates from diverse EDs, reducing biases in the global model. The performance of Hier-FedDRL is evaluated in a cloud-based setup, where Docker containers are used to simulate EDs and Google Kubernetes Engine clusters for cloud orchestration. Experimental results over benchmark datasets demonstrate that the proposed Hier-FedDRL achieves 4%–6% higher accuracy, reduces convergence time by 7%–10%, and lowers bias in the global model by 25%, outperforming state-of-the-art FL approaches while effectively addressing data and resource heterogeneity.
期刊介绍:
The Journal of Network and Computer Applications welcomes research contributions, surveys, and notes in all areas relating to computer networks and applications thereof. Sample topics include new design techniques, interesting or novel applications, components or standards; computer networks with tools such as WWW; emerging standards for internet protocols; Wireless networks; Mobile Computing; emerging computing models such as cloud computing, grid computing; applications of networked systems for remote collaboration and telemedicine, etc. The journal is abstracted and indexed in Scopus, Engineering Index, Web of Science, Science Citation Index Expanded and INSPEC.