{"title":"SHE-SFL:一种基于同态加密的高效且保护隐私的异构联邦分裂学习体系结构","authors":"Jiaqi Xia , Meng Wu , Pengyong Li","doi":"10.1016/j.future.2025.108101","DOIUrl":null,"url":null,"abstract":"<div><div>Federated Learning (FL) and Split Learning (SL) are distributed methods that enable collaborative model training without sharing raw data. Combining FL and SL leverages the benefits of computational offloading and model privacy while enhancing efficiency through parallel processing. However, both methods risk privacy leaks: FL is vulnerable to model inversion attacks and gradient leaks, whereas SL’s data transmission during training can reveal sensitive information, potentially allowing attackers to reconstruct the original dataset. Current privacy protections often fall short of fully securing these systems and impose substantial computational and communication costs. In this work, we introduce SHE-SFL, an efficient privacy-preserving federated split learning architecture based on fully homomorphic encryption. Specifically, we employ the CKKS scheme to encrypt activation values during forward propagation, gradients during backpropagation, and the model parameters shared at the aggregation stage. This ensures that all data leaving the client domain is encrypted. This architecture includes two key modules: SHE-SL encrypts and transmits ciphertext based on batch packing and adopts a sparsification strategy, reducing system overhead and enabling polymorphic training of the models. SHE-Aggr enhances the efficiency of encrypting model parameters during the aggregation phase and perfectly supports encrypted weighted aggregation. Extensive experimental results demonstrate that the proposed SHE-SFL provides comprehensive protection for the federated split learning architecture with minimal impact on model performance, effectively safeguarding client privacy while significantly reducing system overhead.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"175 ","pages":"Article 108101"},"PeriodicalIF":6.2000,"publicationDate":"2025-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"SHE-SFL: An efficient and privacy-preserving heterogeneous federated split learning architecture based on homomorphic encryption\",\"authors\":\"Jiaqi Xia , Meng Wu , Pengyong Li\",\"doi\":\"10.1016/j.future.2025.108101\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Federated Learning (FL) and Split Learning (SL) are distributed methods that enable collaborative model training without sharing raw data. Combining FL and SL leverages the benefits of computational offloading and model privacy while enhancing efficiency through parallel processing. However, both methods risk privacy leaks: FL is vulnerable to model inversion attacks and gradient leaks, whereas SL’s data transmission during training can reveal sensitive information, potentially allowing attackers to reconstruct the original dataset. Current privacy protections often fall short of fully securing these systems and impose substantial computational and communication costs. In this work, we introduce SHE-SFL, an efficient privacy-preserving federated split learning architecture based on fully homomorphic encryption. Specifically, we employ the CKKS scheme to encrypt activation values during forward propagation, gradients during backpropagation, and the model parameters shared at the aggregation stage. This ensures that all data leaving the client domain is encrypted. This architecture includes two key modules: SHE-SL encrypts and transmits ciphertext based on batch packing and adopts a sparsification strategy, reducing system overhead and enabling polymorphic training of the models. SHE-Aggr enhances the efficiency of encrypting model parameters during the aggregation phase and perfectly supports encrypted weighted aggregation. Extensive experimental results demonstrate that the proposed SHE-SFL provides comprehensive protection for the federated split learning architecture with minimal impact on model performance, effectively safeguarding client privacy while significantly reducing system overhead.</div></div>\",\"PeriodicalId\":55132,\"journal\":{\"name\":\"Future Generation Computer Systems-The International Journal of Escience\",\"volume\":\"175 \",\"pages\":\"Article 108101\"},\"PeriodicalIF\":6.2000,\"publicationDate\":\"2025-08-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Future Generation Computer Systems-The International Journal of Escience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167739X25003954\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X25003954","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
SHE-SFL: An efficient and privacy-preserving heterogeneous federated split learning architecture based on homomorphic encryption
Federated Learning (FL) and Split Learning (SL) are distributed methods that enable collaborative model training without sharing raw data. Combining FL and SL leverages the benefits of computational offloading and model privacy while enhancing efficiency through parallel processing. However, both methods risk privacy leaks: FL is vulnerable to model inversion attacks and gradient leaks, whereas SL’s data transmission during training can reveal sensitive information, potentially allowing attackers to reconstruct the original dataset. Current privacy protections often fall short of fully securing these systems and impose substantial computational and communication costs. In this work, we introduce SHE-SFL, an efficient privacy-preserving federated split learning architecture based on fully homomorphic encryption. Specifically, we employ the CKKS scheme to encrypt activation values during forward propagation, gradients during backpropagation, and the model parameters shared at the aggregation stage. This ensures that all data leaving the client domain is encrypted. This architecture includes two key modules: SHE-SL encrypts and transmits ciphertext based on batch packing and adopts a sparsification strategy, reducing system overhead and enabling polymorphic training of the models. SHE-Aggr enhances the efficiency of encrypting model parameters during the aggregation phase and perfectly supports encrypted weighted aggregation. Extensive experimental results demonstrate that the proposed SHE-SFL provides comprehensive protection for the federated split learning architecture with minimal impact on model performance, effectively safeguarding client privacy while significantly reducing system overhead.
期刊介绍:
Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications.
Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration.
Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.