Zhenyong Zhang , Kuan Shao , Ruilong Deng , Xin Wang , Yishu Zhang , Mufeng Wang
{"title":"PrivLSTM:融合了加密和网络结构的多源数据隐私保护LSTM推理框架","authors":"Zhenyong Zhang , Kuan Shao , Ruilong Deng , Xin Wang , Yishu Zhang , Mufeng Wang","doi":"10.1016/j.inffus.2025.103711","DOIUrl":null,"url":null,"abstract":"<div><div>Machine Learning as a Service (MLaaS) has emerged as a prominent topic for dealing with multi-sourced data. However, the privacy concerns have received more considerable attention than ever. In this paper, we propose PrivLSTM, a low latency privacy-preserving long short-term memory (LSTM) neural network inference method by fusing the encrypted data and network structure, protecting the sequence data input to the LSTM. PrivLSTM is implemented with a non-interactive framework using fully homomorphic encryption and is adaptive to a general LSTM structure with long input sequences. PrivLSTM uses bootstrapping to avoid the impact of noise introduced by the deep multiplications on the accuracy. The computation overhead of bootstrapping is alleviated by developing a novel batch-based linear transformation method, which reduces the automorphism operations and integrates identical operations in the LSTM cell. The nonlinear operations, e.g., sigmoid and tanh, are approximated with optimized polynomials. Besides, the efficiency and security of PrivLSTM are theoretically analyzed. We conduct experiments on multi-sourced datasets, such as text and sensory datasets. The results show that PrivLSTM achieves comparable accuracy against the plain LSTM and fast computation with around 30s for a 100-length input sequence.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"127 ","pages":"Article 103711"},"PeriodicalIF":15.5000,"publicationDate":"2025-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"PrivLSTM: A privacy-preserving LSTM inference framework by fusing encryption and network structure for multi-sourced Data\",\"authors\":\"Zhenyong Zhang , Kuan Shao , Ruilong Deng , Xin Wang , Yishu Zhang , Mufeng Wang\",\"doi\":\"10.1016/j.inffus.2025.103711\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Machine Learning as a Service (MLaaS) has emerged as a prominent topic for dealing with multi-sourced data. However, the privacy concerns have received more considerable attention than ever. In this paper, we propose PrivLSTM, a low latency privacy-preserving long short-term memory (LSTM) neural network inference method by fusing the encrypted data and network structure, protecting the sequence data input to the LSTM. PrivLSTM is implemented with a non-interactive framework using fully homomorphic encryption and is adaptive to a general LSTM structure with long input sequences. PrivLSTM uses bootstrapping to avoid the impact of noise introduced by the deep multiplications on the accuracy. The computation overhead of bootstrapping is alleviated by developing a novel batch-based linear transformation method, which reduces the automorphism operations and integrates identical operations in the LSTM cell. The nonlinear operations, e.g., sigmoid and tanh, are approximated with optimized polynomials. Besides, the efficiency and security of PrivLSTM are theoretically analyzed. We conduct experiments on multi-sourced datasets, such as text and sensory datasets. The results show that PrivLSTM achieves comparable accuracy against the plain LSTM and fast computation with around 30s for a 100-length input sequence.</div></div>\",\"PeriodicalId\":50367,\"journal\":{\"name\":\"Information Fusion\",\"volume\":\"127 \",\"pages\":\"Article 103711\"},\"PeriodicalIF\":15.5000,\"publicationDate\":\"2025-09-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Fusion\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1566253525007444\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525007444","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
PrivLSTM: A privacy-preserving LSTM inference framework by fusing encryption and network structure for multi-sourced Data
Machine Learning as a Service (MLaaS) has emerged as a prominent topic for dealing with multi-sourced data. However, the privacy concerns have received more considerable attention than ever. In this paper, we propose PrivLSTM, a low latency privacy-preserving long short-term memory (LSTM) neural network inference method by fusing the encrypted data and network structure, protecting the sequence data input to the LSTM. PrivLSTM is implemented with a non-interactive framework using fully homomorphic encryption and is adaptive to a general LSTM structure with long input sequences. PrivLSTM uses bootstrapping to avoid the impact of noise introduced by the deep multiplications on the accuracy. The computation overhead of bootstrapping is alleviated by developing a novel batch-based linear transformation method, which reduces the automorphism operations and integrates identical operations in the LSTM cell. The nonlinear operations, e.g., sigmoid and tanh, are approximated with optimized polynomials. Besides, the efficiency and security of PrivLSTM are theoretically analyzed. We conduct experiments on multi-sourced datasets, such as text and sensory datasets. The results show that PrivLSTM achieves comparable accuracy against the plain LSTM and fast computation with around 30s for a 100-length input sequence.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.