{"title":"类感知的高效客户端选择和多损失引导下的联邦学习局部性能优化","authors":"Akshay Singh, Rahul Thakur","doi":"10.1016/j.future.2025.108172","DOIUrl":null,"url":null,"abstract":"<div><div>Federated Learning (FL) enables decentralized model training across multiple edge devices, preserving privacy through local data processing at the edge. While FL reduces data transfer by exchanging only model weights, inherent multiple aggregation steps can increase the communication overhead. Practically, clients’ datasets exhibit significant variation in their label distributions. This means that data across clients is not independently and identically distributed (non-IID) with respect to label classes, creating distinct patterns unique to each client. It causes inconsistencies between local objectives and global optima, which impacts the overall performance of global training. Additionally, trade-offs between performance and energy consumption remain a key challenge mostly ignored by existing FL methods. To address this issue, we introduce a class distribution-aware client selection algorithm guided by a multi-loss function that optimizes both performance and energy consumption across participating edge devices, named CACS-FL. In CACS-FL, firstly, a confidence score is calculated for each edge device based on their heterogeneity score and energy consumption, which is further optimized to select an appropriate set of edge devices. After selection, the clients perform local training using a weighted multi-loss function, which improves personalized performance and achieves higher global performance. Experimental demonstrations on various datasets showcase CACS-FL’s advantages over existing state-of-the-art approaches. CACS-FL also guarantees faster convergence with fairness in performance across the participating edge devices.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"176 ","pages":"Article 108172"},"PeriodicalIF":6.2000,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Class aware efficient client selection and multi-loss guided local performance optimization in federated learning\",\"authors\":\"Akshay Singh, Rahul Thakur\",\"doi\":\"10.1016/j.future.2025.108172\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Federated Learning (FL) enables decentralized model training across multiple edge devices, preserving privacy through local data processing at the edge. While FL reduces data transfer by exchanging only model weights, inherent multiple aggregation steps can increase the communication overhead. Practically, clients’ datasets exhibit significant variation in their label distributions. This means that data across clients is not independently and identically distributed (non-IID) with respect to label classes, creating distinct patterns unique to each client. It causes inconsistencies between local objectives and global optima, which impacts the overall performance of global training. Additionally, trade-offs between performance and energy consumption remain a key challenge mostly ignored by existing FL methods. To address this issue, we introduce a class distribution-aware client selection algorithm guided by a multi-loss function that optimizes both performance and energy consumption across participating edge devices, named CACS-FL. In CACS-FL, firstly, a confidence score is calculated for each edge device based on their heterogeneity score and energy consumption, which is further optimized to select an appropriate set of edge devices. After selection, the clients perform local training using a weighted multi-loss function, which improves personalized performance and achieves higher global performance. Experimental demonstrations on various datasets showcase CACS-FL’s advantages over existing state-of-the-art approaches. CACS-FL also guarantees faster convergence with fairness in performance across the participating edge devices.</div></div>\",\"PeriodicalId\":55132,\"journal\":{\"name\":\"Future Generation Computer Systems-The International Journal of Escience\",\"volume\":\"176 \",\"pages\":\"Article 108172\"},\"PeriodicalIF\":6.2000,\"publicationDate\":\"2025-10-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Future Generation Computer Systems-The International Journal of Escience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167739X25004662\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X25004662","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
Class aware efficient client selection and multi-loss guided local performance optimization in federated learning
Federated Learning (FL) enables decentralized model training across multiple edge devices, preserving privacy through local data processing at the edge. While FL reduces data transfer by exchanging only model weights, inherent multiple aggregation steps can increase the communication overhead. Practically, clients’ datasets exhibit significant variation in their label distributions. This means that data across clients is not independently and identically distributed (non-IID) with respect to label classes, creating distinct patterns unique to each client. It causes inconsistencies between local objectives and global optima, which impacts the overall performance of global training. Additionally, trade-offs between performance and energy consumption remain a key challenge mostly ignored by existing FL methods. To address this issue, we introduce a class distribution-aware client selection algorithm guided by a multi-loss function that optimizes both performance and energy consumption across participating edge devices, named CACS-FL. In CACS-FL, firstly, a confidence score is calculated for each edge device based on their heterogeneity score and energy consumption, which is further optimized to select an appropriate set of edge devices. After selection, the clients perform local training using a weighted multi-loss function, which improves personalized performance and achieves higher global performance. Experimental demonstrations on various datasets showcase CACS-FL’s advantages over existing state-of-the-art approaches. CACS-FL also guarantees faster convergence with fairness in performance across the participating edge devices.
期刊介绍:
Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications.
Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration.
Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.