Multi-View Federated Learning with Data Collaboration

Yitao Yang, Xiucai Ye, Tetsuya Sakurai
{"title":"Multi-View Federated Learning with Data Collaboration","authors":"Yitao Yang, Xiucai Ye, Tetsuya Sakurai","doi":"10.1145/3529836.3529904","DOIUrl":null,"url":null,"abstract":"Under the privacy protection policy, federated learning has received more and more attention. Vertical federated learning (VFL) uses the same samples local in different parties to build prediction model. However, the same samples (overlapping samples) may be limited, while a large number of non-overlapping samples in each party are not utilized. If the non-overlapping samples can be utilized for training, it can benefit the prediction model. In this paper, we propose a novel VFL method, called Multi-View Federated Learning with Data collaboration (FedMC), to solve the problem of insufficient overlapping samples by exploiting suitable non-overlapping samples for data training. The proposed FedMC method first constructs a common feature space based on the overlapping samples, then projects the non-overlapping samples into the common feature space. We measure the similarity for each pair of the non-overlapping samples by calculating their distance in this space. When the distance is less than a threshold, we match them and add this pair to the overlapping samples. The expanded overlapping samples are finally used for training to build the prediction model. We evaluate the proposed method on real-world datasets. The experimental results show that the proposed method can improve the classification result by exploiting the non-overlapping samples for training.","PeriodicalId":285191,"journal":{"name":"2022 14th International Conference on Machine Learning and Computing (ICMLC)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 14th International Conference on Machine Learning and Computing (ICMLC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3529836.3529904","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5

Abstract

Under the privacy protection policy, federated learning has received more and more attention. Vertical federated learning (VFL) uses the same samples local in different parties to build prediction model. However, the same samples (overlapping samples) may be limited, while a large number of non-overlapping samples in each party are not utilized. If the non-overlapping samples can be utilized for training, it can benefit the prediction model. In this paper, we propose a novel VFL method, called Multi-View Federated Learning with Data collaboration (FedMC), to solve the problem of insufficient overlapping samples by exploiting suitable non-overlapping samples for data training. The proposed FedMC method first constructs a common feature space based on the overlapping samples, then projects the non-overlapping samples into the common feature space. We measure the similarity for each pair of the non-overlapping samples by calculating their distance in this space. When the distance is less than a threshold, we match them and add this pair to the overlapping samples. The expanded overlapping samples are finally used for training to build the prediction model. We evaluate the proposed method on real-world datasets. The experimental results show that the proposed method can improve the classification result by exploiting the non-overlapping samples for training.
数据协作的多视图联邦学习
在隐私保护政策下,联邦学习受到越来越多的关注。垂直联邦学习(Vertical federated learning, VFL)利用本地不同方的相同样本构建预测模型。但是,相同的样本(重叠的样本)可能是有限的,而每一方的大量不重叠的样本没有被利用。如果能利用不重叠的样本进行训练,对预测模型是有利的。在本文中,我们提出了一种新的VFL方法,称为多视图联邦学习与数据协作(FedMC),通过利用合适的非重叠样本进行数据训练来解决重叠样本不足的问题。提出的FedMC方法首先基于重叠样本构建一个公共特征空间,然后将非重叠样本投影到公共特征空间中。我们通过计算它们在这个空间中的距离来测量每对非重叠样本的相似性。当距离小于阈值时,我们将它们匹配并将这对添加到重叠的样本中。最后利用扩展后的重叠样本进行训练,建立预测模型。我们在真实世界的数据集上评估了所提出的方法。实验结果表明,该方法利用非重叠样本进行训练,可以提高分类效果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信