Van An Le, Jason Haga, Yusuke Tanimura, Truong Thao Nguyen
{"title":"FLaTEC: An efficient federated learning scheme across the Thing-Edge-Cloud environment","authors":"Van An Le, Jason Haga, Yusuke Tanimura, Truong Thao Nguyen","doi":"10.1016/j.future.2025.108073","DOIUrl":null,"url":null,"abstract":"<div><div>Federated Learning (FL) has become a cornerstone for enabling decentralized model training in mobile edge computing and Internet of Things (IoT) environments and maintaining data privacy by keeping data local to devices. However, the exponential growth in the number of participating devices and the increasing size and complexity of Machine Learning (ML) models amplify FL’s challenges, including high communication overhead, significant computational and energy constraints on edge devices, and the issue of heterogeneous data distribution, i.e., non-Independent and Identically Distributed (non-IID) data across clients. To address these challenges, we propose FLaTEC, a novel FL system tailored for the Thing-Edge-Cloud (TEC) architecture. First, FLaTEC introduces a split-training architecture that divides the global model into three components: a lightweight base model trained on resource-constrained edge devices, a computationally intensive core model trained on edge servers, and a simplified core model designed for on-device training and inference. Second, FLaTEC adopts a separate training strategy in which feature data is uploaded periodically from devices to edge servers to train the core model, reducing frequent data exchanges and mitigating the non-IID problem in FL. Third, to enhance the performance of the simplified model used for on-device training, FLaTEC applies knowledge distillation from the core model trained at the edge. A cloud server orchestrates the entire system by aggregating the base, core, and simplified core models using a federated averaging algorithm, ensuring consistency and coordination across devices and edge servers. Extensive experiments conducted across multiple datasets and diverse ML tasks validate FLaTEC’s superior performance, demonstrating its ability to achieve high accuracy, reduced communication overhead, and resilience to data heterogeneity compared to state-of-the-art methods.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"175 ","pages":"Article 108073"},"PeriodicalIF":6.2000,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X2500367X","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
Federated Learning (FL) has become a cornerstone for enabling decentralized model training in mobile edge computing and Internet of Things (IoT) environments and maintaining data privacy by keeping data local to devices. However, the exponential growth in the number of participating devices and the increasing size and complexity of Machine Learning (ML) models amplify FL’s challenges, including high communication overhead, significant computational and energy constraints on edge devices, and the issue of heterogeneous data distribution, i.e., non-Independent and Identically Distributed (non-IID) data across clients. To address these challenges, we propose FLaTEC, a novel FL system tailored for the Thing-Edge-Cloud (TEC) architecture. First, FLaTEC introduces a split-training architecture that divides the global model into three components: a lightweight base model trained on resource-constrained edge devices, a computationally intensive core model trained on edge servers, and a simplified core model designed for on-device training and inference. Second, FLaTEC adopts a separate training strategy in which feature data is uploaded periodically from devices to edge servers to train the core model, reducing frequent data exchanges and mitigating the non-IID problem in FL. Third, to enhance the performance of the simplified model used for on-device training, FLaTEC applies knowledge distillation from the core model trained at the edge. A cloud server orchestrates the entire system by aggregating the base, core, and simplified core models using a federated averaging algorithm, ensuring consistency and coordination across devices and edge servers. Extensive experiments conducted across multiple datasets and diverse ML tasks validate FLaTEC’s superior performance, demonstrating its ability to achieve high accuracy, reduced communication overhead, and resilience to data heterogeneity compared to state-of-the-art methods.
期刊介绍:
Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications.
Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration.
Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.