{"title":"隐私保护联邦学习的私有数据预处理","authors":"Ruei-Hau Hsu, Ting-Yun Huang","doi":"10.1109/ICKII55100.2022.9983518","DOIUrl":null,"url":null,"abstract":"Privacy-preserving federated learning can accomplish model aggregation without leaking to a local model and avoid the problem of sensitive data exposure caused by model leakage. Even though it protects privacy in the training process, any data analysis task proposed by the initiator and the type of data required for the task contains the research or trade secrets of the organization. Being intercepted in the transmission process or known by other data providers, the disclosure of essential research secrets occurs, leading to the theft of research or business ideas. Therefore, it is a critical issue to achieve data matching between the initiator and the participant under the premise of privacy protection. In this study, we propose a federated learning framework that considers the above security issues. A privacy-preserving federated learning architecture based on homomorphic encryption is designed to protect each participant's data and local model. In addition, encrypted query technology is used in this architecture to provide data privacy matching. The data provider searches the data in ciphertext, finds the encrypted data that meets the conditions, and completes the training process without disclosing any requirements of the task initiator.","PeriodicalId":352222,"journal":{"name":"2022 IEEE 5th International Conference on Knowledge Innovation and Invention (ICKII )","volume":"49 18","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Private Data Preprocessing for Privacy-preserving Federated Learning\",\"authors\":\"Ruei-Hau Hsu, Ting-Yun Huang\",\"doi\":\"10.1109/ICKII55100.2022.9983518\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Privacy-preserving federated learning can accomplish model aggregation without leaking to a local model and avoid the problem of sensitive data exposure caused by model leakage. Even though it protects privacy in the training process, any data analysis task proposed by the initiator and the type of data required for the task contains the research or trade secrets of the organization. Being intercepted in the transmission process or known by other data providers, the disclosure of essential research secrets occurs, leading to the theft of research or business ideas. Therefore, it is a critical issue to achieve data matching between the initiator and the participant under the premise of privacy protection. In this study, we propose a federated learning framework that considers the above security issues. A privacy-preserving federated learning architecture based on homomorphic encryption is designed to protect each participant's data and local model. In addition, encrypted query technology is used in this architecture to provide data privacy matching. The data provider searches the data in ciphertext, finds the encrypted data that meets the conditions, and completes the training process without disclosing any requirements of the task initiator.\",\"PeriodicalId\":352222,\"journal\":{\"name\":\"2022 IEEE 5th International Conference on Knowledge Innovation and Invention (ICKII )\",\"volume\":\"49 18\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-07-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 5th International Conference on Knowledge Innovation and Invention (ICKII )\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICKII55100.2022.9983518\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 5th International Conference on Knowledge Innovation and Invention (ICKII )","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICKII55100.2022.9983518","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Private Data Preprocessing for Privacy-preserving Federated Learning
Privacy-preserving federated learning can accomplish model aggregation without leaking to a local model and avoid the problem of sensitive data exposure caused by model leakage. Even though it protects privacy in the training process, any data analysis task proposed by the initiator and the type of data required for the task contains the research or trade secrets of the organization. Being intercepted in the transmission process or known by other data providers, the disclosure of essential research secrets occurs, leading to the theft of research or business ideas. Therefore, it is a critical issue to achieve data matching between the initiator and the participant under the premise of privacy protection. In this study, we propose a federated learning framework that considers the above security issues. A privacy-preserving federated learning architecture based on homomorphic encryption is designed to protect each participant's data and local model. In addition, encrypted query technology is used in this architecture to provide data privacy matching. The data provider searches the data in ciphertext, finds the encrypted data that meets the conditions, and completes the training process without disclosing any requirements of the task initiator.