{"title":"Resource-Efficient Federated Clustering with Past Negatives Pool","authors":"Runxuan Miao, Erdem Koyuncu","doi":"10.1109/ICASSPW59220.2023.10193449","DOIUrl":null,"url":null,"abstract":"Federated learning (FL) provides a global model over data distributed to multiple clients. However, most recent work on FL focuses on supervised learning, and a fully unsupervised federated clustering scheme has remained an open problem. In this context, Contrastive learning (CL) trains distinguishable instance embeddings without labels. However, most CL techniques are restricted to centralized data. In this work, we consider the problem of clustering data that is distributed to multiple clients using FL and CL. We propose a federated clustering framework with a novel past negatives pool (PNP) for intelligently selecting positive and negative samples for CL. PNP benefits FL and CL simultaneously, specifically, alleviating class collision for CL and reducing client-drift in FL. PNP thus provides a higher accuracy for a given constraint on the communication rounds, which makes it suitable for networks with limited communication and computation resources. Numerical results show that the resulting FedPNP scheme achieves superior performance in solving federated clustering problems on benchmark datasets including CIFAR-10 and CIFAR-100, especially in non-iid settings.","PeriodicalId":158726,"journal":{"name":"2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICASSPW59220.2023.10193449","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Federated learning (FL) provides a global model over data distributed to multiple clients. However, most recent work on FL focuses on supervised learning, and a fully unsupervised federated clustering scheme has remained an open problem. In this context, Contrastive learning (CL) trains distinguishable instance embeddings without labels. However, most CL techniques are restricted to centralized data. In this work, we consider the problem of clustering data that is distributed to multiple clients using FL and CL. We propose a federated clustering framework with a novel past negatives pool (PNP) for intelligently selecting positive and negative samples for CL. PNP benefits FL and CL simultaneously, specifically, alleviating class collision for CL and reducing client-drift in FL. PNP thus provides a higher accuracy for a given constraint on the communication rounds, which makes it suitable for networks with limited communication and computation resources. Numerical results show that the resulting FedPNP scheme achieves superior performance in solving federated clustering problems on benchmark datasets including CIFAR-10 and CIFAR-100, especially in non-iid settings.