{"title":"TPFL:基于置信度聚类的蔡特林个性化联合学习","authors":"Rasoul Jafari Gohari, Laya Aliahmadipour, Ezat Valipour","doi":"arxiv-2409.10392","DOIUrl":null,"url":null,"abstract":"The world of Machine Learning (ML) has witnessed rapid changes in terms of\nnew models and ways to process users data. The majority of work that has been\ndone is focused on Deep Learning (DL) based approaches. However, with the\nemergence of new algorithms such as the Tsetlin Machine (TM) algorithm, there\nis growing interest in exploring alternative approaches that may offer unique\nadvantages in certain domains or applications. One of these domains is\nFederated Learning (FL), in which users privacy is of utmost importance. Due to\nits novelty, FL has seen a surge in the incorporation of personalization\ntechniques to enhance model accuracy while maintaining user privacy under\npersonalized conditions. In this work, we propose a novel approach dubbed TPFL:\nTsetlin-Personalized Federated Learning, in which models are grouped into\nclusters based on their confidence towards a specific class. In this way,\nclustering can benefit from two key advantages. Firstly, clients share only\nwhat they are confident about, resulting in the elimination of wrongful weight\naggregation among clients whose data for a specific class may have not been\nenough during the training. This phenomenon is prevalent when the data are\nnon-Independent and Identically Distributed (non-IID). Secondly, by sharing\nonly weights towards a specific class, communication cost is substantially\nreduced, making TPLF efficient in terms of both accuracy and communication\ncost. The results of TPFL demonstrated the highest accuracy on three different\ndatasets; namely MNIST, FashionMNIST and FEMNIST.","PeriodicalId":501422,"journal":{"name":"arXiv - CS - Distributed, Parallel, and Cluster Computing","volume":"52 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"TPFL: Tsetlin-Personalized Federated Learning with Confidence-Based Clustering\",\"authors\":\"Rasoul Jafari Gohari, Laya Aliahmadipour, Ezat Valipour\",\"doi\":\"arxiv-2409.10392\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The world of Machine Learning (ML) has witnessed rapid changes in terms of\\nnew models and ways to process users data. The majority of work that has been\\ndone is focused on Deep Learning (DL) based approaches. However, with the\\nemergence of new algorithms such as the Tsetlin Machine (TM) algorithm, there\\nis growing interest in exploring alternative approaches that may offer unique\\nadvantages in certain domains or applications. One of these domains is\\nFederated Learning (FL), in which users privacy is of utmost importance. Due to\\nits novelty, FL has seen a surge in the incorporation of personalization\\ntechniques to enhance model accuracy while maintaining user privacy under\\npersonalized conditions. In this work, we propose a novel approach dubbed TPFL:\\nTsetlin-Personalized Federated Learning, in which models are grouped into\\nclusters based on their confidence towards a specific class. In this way,\\nclustering can benefit from two key advantages. Firstly, clients share only\\nwhat they are confident about, resulting in the elimination of wrongful weight\\naggregation among clients whose data for a specific class may have not been\\nenough during the training. This phenomenon is prevalent when the data are\\nnon-Independent and Identically Distributed (non-IID). Secondly, by sharing\\nonly weights towards a specific class, communication cost is substantially\\nreduced, making TPLF efficient in terms of both accuracy and communication\\ncost. The results of TPFL demonstrated the highest accuracy on three different\\ndatasets; namely MNIST, FashionMNIST and FEMNIST.\",\"PeriodicalId\":501422,\"journal\":{\"name\":\"arXiv - CS - Distributed, Parallel, and Cluster Computing\",\"volume\":\"52 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Distributed, Parallel, and Cluster Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.10392\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Distributed, Parallel, and Cluster Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10392","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
TPFL: Tsetlin-Personalized Federated Learning with Confidence-Based Clustering
The world of Machine Learning (ML) has witnessed rapid changes in terms of
new models and ways to process users data. The majority of work that has been
done is focused on Deep Learning (DL) based approaches. However, with the
emergence of new algorithms such as the Tsetlin Machine (TM) algorithm, there
is growing interest in exploring alternative approaches that may offer unique
advantages in certain domains or applications. One of these domains is
Federated Learning (FL), in which users privacy is of utmost importance. Due to
its novelty, FL has seen a surge in the incorporation of personalization
techniques to enhance model accuracy while maintaining user privacy under
personalized conditions. In this work, we propose a novel approach dubbed TPFL:
Tsetlin-Personalized Federated Learning, in which models are grouped into
clusters based on their confidence towards a specific class. In this way,
clustering can benefit from two key advantages. Firstly, clients share only
what they are confident about, resulting in the elimination of wrongful weight
aggregation among clients whose data for a specific class may have not been
enough during the training. This phenomenon is prevalent when the data are
non-Independent and Identically Distributed (non-IID). Secondly, by sharing
only weights towards a specific class, communication cost is substantially
reduced, making TPLF efficient in terms of both accuracy and communication
cost. The results of TPFL demonstrated the highest accuracy on three different
datasets; namely MNIST, FashionMNIST and FEMNIST.