PerHeFed:异构卷积神经网络个性化联合学习通用框架

IF 2.7 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS
Le Ma, YuYing Liao, Bin Zhou, Wen Xi
{"title":"PerHeFed:异构卷积神经网络个性化联合学习通用框架","authors":"Le Ma, YuYing Liao, Bin Zhou, Wen Xi","doi":"10.1007/s11280-022-01119-x","DOIUrl":null,"url":null,"abstract":"<p><p>In conventional federated learning, each device is restricted to train a network model of the same structure. This greatly hinders the application of federated learning where the data and devices are quite heterogeneous because of their different hardware equipment and communication networks. At the same time, existing studies have shown that transmitting all of the model parameters not only has heavy communication costs, but also increases risk of privacy leakage. We propose a general framework for personalized federated learning (PerHeFed), which enables the devices to design their local model structures autonomously and share sub-models without structural restrictions. In PerHeFed, a simple-but-effective mapping relation and a novel personalized sub-model aggregation method are proposed for heterogeneous sub-models to be aggregated. By dividing the aggregations into two primitive types (i.e., inter-layer and intra-layer), PerHeFed is applicable to any combination of heterogeneous convolutional neural networks, and we believe that this can satisfy the personalized requirements of heterogeneous models. Experiments show that, compared to the state-of-the-art method (e.g., FLOP), in non-IID data sets our method compress ≈ 50<i>%</i> of the shared sub-model parameters with only a 4.38% drop in accuracy on SVHN dataset and on CIFAR-10, PerHeFed even achieves a 0.3% improvement in accuracy. To the best of our knowledge, our work is the first general personalized federated learning framework for heterogeneous convolutional networks, even cross different networks, addressing model structure unity in conventional federated learning.</p>","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":null,"pages":null},"PeriodicalIF":2.7000,"publicationDate":"2022-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9743105/pdf/","citationCount":"0","resultStr":"{\"title\":\"PerHeFed: A general framework of personalized federated learning for heterogeneous convolutional neural networks.\",\"authors\":\"Le Ma, YuYing Liao, Bin Zhou, Wen Xi\",\"doi\":\"10.1007/s11280-022-01119-x\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>In conventional federated learning, each device is restricted to train a network model of the same structure. This greatly hinders the application of federated learning where the data and devices are quite heterogeneous because of their different hardware equipment and communication networks. At the same time, existing studies have shown that transmitting all of the model parameters not only has heavy communication costs, but also increases risk of privacy leakage. We propose a general framework for personalized federated learning (PerHeFed), which enables the devices to design their local model structures autonomously and share sub-models without structural restrictions. In PerHeFed, a simple-but-effective mapping relation and a novel personalized sub-model aggregation method are proposed for heterogeneous sub-models to be aggregated. By dividing the aggregations into two primitive types (i.e., inter-layer and intra-layer), PerHeFed is applicable to any combination of heterogeneous convolutional neural networks, and we believe that this can satisfy the personalized requirements of heterogeneous models. Experiments show that, compared to the state-of-the-art method (e.g., FLOP), in non-IID data sets our method compress ≈ 50<i>%</i> of the shared sub-model parameters with only a 4.38% drop in accuracy on SVHN dataset and on CIFAR-10, PerHeFed even achieves a 0.3% improvement in accuracy. To the best of our knowledge, our work is the first general personalized federated learning framework for heterogeneous convolutional networks, even cross different networks, addressing model structure unity in conventional federated learning.</p>\",\"PeriodicalId\":49356,\"journal\":{\"name\":\"World Wide Web-Internet and Web Information Systems\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":2.7000,\"publicationDate\":\"2022-12-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9743105/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"World Wide Web-Internet and Web Information Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.1007/s11280-022-01119-x\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"World Wide Web-Internet and Web Information Systems","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s11280-022-01119-x","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

摘要

在传统的联合学习中,每个设备只能训练一个结构相同的网络模型。这极大地阻碍了联合学习的应用,因为在联合学习中,数据和设备因硬件设备和通信网络的不同而具有很大的异质性。同时,现有研究表明,传输所有模型参数不仅通信成本高昂,还会增加隐私泄露的风险。我们提出了一种用于个性化联合学习的通用框架(PerHeFed),它能让设备自主设计本地模型结构,并不受结构限制地共享子模型。在 PerHeFed 中,我们提出了一种简单而有效的映射关系和一种新颖的个性化子模型聚合方法,用于聚合异构子模型。通过将聚合分为两种原始类型(即层间和层内),PerHeFed 适用于任何异构卷积神经网络的组合,我们相信这可以满足异构模型的个性化要求。实验表明,与最先进的方法(如 FLOP)相比,在非 IID 数据集上,我们的方法压缩了≈50% 的共享子模型参数,而在 SVHN 数据集上,准确率仅下降了 4.38%;在 CIFAR-10 数据集上,PerHeFed 甚至提高了 0.3% 的准确率。据我们所知,我们的工作是第一个针对异构卷积网络(甚至是跨不同网络)的通用个性化联合学习框架,解决了传统联合学习中模型结构不统一的问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

PerHeFed: A general framework of personalized federated learning for heterogeneous convolutional neural networks.

PerHeFed: A general framework of personalized federated learning for heterogeneous convolutional neural networks.

PerHeFed: A general framework of personalized federated learning for heterogeneous convolutional neural networks.

PerHeFed: A general framework of personalized federated learning for heterogeneous convolutional neural networks.

In conventional federated learning, each device is restricted to train a network model of the same structure. This greatly hinders the application of federated learning where the data and devices are quite heterogeneous because of their different hardware equipment and communication networks. At the same time, existing studies have shown that transmitting all of the model parameters not only has heavy communication costs, but also increases risk of privacy leakage. We propose a general framework for personalized federated learning (PerHeFed), which enables the devices to design their local model structures autonomously and share sub-models without structural restrictions. In PerHeFed, a simple-but-effective mapping relation and a novel personalized sub-model aggregation method are proposed for heterogeneous sub-models to be aggregated. By dividing the aggregations into two primitive types (i.e., inter-layer and intra-layer), PerHeFed is applicable to any combination of heterogeneous convolutional neural networks, and we believe that this can satisfy the personalized requirements of heterogeneous models. Experiments show that, compared to the state-of-the-art method (e.g., FLOP), in non-IID data sets our method compress ≈ 50% of the shared sub-model parameters with only a 4.38% drop in accuracy on SVHN dataset and on CIFAR-10, PerHeFed even achieves a 0.3% improvement in accuracy. To the best of our knowledge, our work is the first general personalized federated learning framework for heterogeneous convolutional networks, even cross different networks, addressing model structure unity in conventional federated learning.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
World Wide Web-Internet and Web Information Systems
World Wide Web-Internet and Web Information Systems 工程技术-计算机:软件工程
CiteScore
7.30
自引率
10.80%
发文量
131
审稿时长
6 months
期刊介绍: World Wide Web: Internet and Web Information Systems (WWW) is an international, archival, peer-reviewed journal which covers all aspects of the World Wide Web, including issues related to architectures, applications, Internet and Web information systems, and communities. The purpose of this journal is to provide an international forum for researchers, professionals, and industrial practitioners to share their rapidly developing knowledge and report on new advances in Internet and web-based systems. The journal also focuses on all database- and information-system topics that relate to the Internet and the Web, particularly on ways to model, design, develop, integrate, and manage these systems. Appearing quarterly, the journal publishes (1) papers describing original ideas and new results, (2) vision papers, (3) reviews of important techniques in related areas, (4) innovative application papers, and (5) progress reports on major international research projects. Papers published in the WWW journal deal with subjects directly or indirectly related to the World Wide Web. The WWW journal provides timely, in-depth coverage of the most recent developments in the World Wide Web discipline to enable anyone involved to keep up-to-date with this dynamically changing technology.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信