Harshit Gupta , Arya Krishnan , O.P. Vyas , Giovanni Merlino , Francesco Longo , Antonio Puliafito
{"title":"Federated learning across the compute continuum: A hierarchical approach with splitNNs and personalized layers","authors":"Harshit Gupta , Arya Krishnan , O.P. Vyas , Giovanni Merlino , Francesco Longo , Antonio Puliafito","doi":"10.1016/j.future.2025.107878","DOIUrl":null,"url":null,"abstract":"<div><div>Federated Learning (FL) allows a Machine Learning (ML) model to be trained collaboratively among distributed devices while preserving the privacy of the data being used for the training. On the other hand, Hierarchical Federated Learning (HFL) is the extended architecture of FL, which consists of additional edge servers for partial aggregation. FL is very useful in privacy-preserving machine learning. However, it has some setbacks, such as statistical heterogeneity, multiple expensive global iterations, performance degradation due to insufficient data, and slow convergence. To deal with such setbacks, the work proposes three approaches with HFL. The first approach utilizes Transfer Learning with HFL, the second approach uses personalized layers in HFL by presenting a 2-tier & 3-tier architecture, and the third approach uses Split Learning (SL) with HFL by proposing an extended 3-tier architecture. The proposed work performed well with the computation at multilevel, i.e., on client, edge, and cloud, exploiting the hybrid infrastructure of IoT-Edge-cloud, i.e., compute continuum. The obtained results showed that the proposed work outperforms by increasing the accuracy of complex models from 18.10% to 76.91% with faster convergence. The work also showed better performance than the state-of-the-art models. Significant performance improvement was achieved in the presence of personalized layers in an HFL-SplitNN architecture. The proposed 3-tier architecture especially shines in the case of less homogeneous data per client. SL played a vital role with HFL in enhancing performance by providing a maximum accuracy of 82.38% with Independent & Identically Distributed Data (IID) and 52.16% with Non-IID data distribution.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"173 ","pages":"Article 107878"},"PeriodicalIF":6.2000,"publicationDate":"2025-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X25001736","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
Federated Learning (FL) allows a Machine Learning (ML) model to be trained collaboratively among distributed devices while preserving the privacy of the data being used for the training. On the other hand, Hierarchical Federated Learning (HFL) is the extended architecture of FL, which consists of additional edge servers for partial aggregation. FL is very useful in privacy-preserving machine learning. However, it has some setbacks, such as statistical heterogeneity, multiple expensive global iterations, performance degradation due to insufficient data, and slow convergence. To deal with such setbacks, the work proposes three approaches with HFL. The first approach utilizes Transfer Learning with HFL, the second approach uses personalized layers in HFL by presenting a 2-tier & 3-tier architecture, and the third approach uses Split Learning (SL) with HFL by proposing an extended 3-tier architecture. The proposed work performed well with the computation at multilevel, i.e., on client, edge, and cloud, exploiting the hybrid infrastructure of IoT-Edge-cloud, i.e., compute continuum. The obtained results showed that the proposed work outperforms by increasing the accuracy of complex models from 18.10% to 76.91% with faster convergence. The work also showed better performance than the state-of-the-art models. Significant performance improvement was achieved in the presence of personalized layers in an HFL-SplitNN architecture. The proposed 3-tier architecture especially shines in the case of less homogeneous data per client. SL played a vital role with HFL in enhancing performance by providing a maximum accuracy of 82.38% with Independent & Identically Distributed Data (IID) and 52.16% with Non-IID data distribution.
期刊介绍:
Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications.
Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration.
Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.