Aristeidis Karras, Christos N. Karras, K. Giotopoulos, D. Tsolis, K. Oikonomou, S. Sioutas
{"title":"Peer to Peer Federated Learning: Towards Decentralized Machine Learning on Edge Devices","authors":"Aristeidis Karras, Christos N. Karras, K. Giotopoulos, D. Tsolis, K. Oikonomou, S. Sioutas","doi":"10.1109/SEEDA-CECNSM57760.2022.9932980","DOIUrl":null,"url":null,"abstract":"Federated Learning (FL) is an emerging technique that assures user privacy and data integrity in distributed machine learning environments. To perform so, chunks of data are trained across edge devices and a high performance cluster server maintains a local copy without exchanging it with other parties. In this work, we investigate a FL scenario in a real-world case study using 5, 10 and 20 Raspberry Pi devices acting as clients. Under this setup, we employ the widely known FedAvg algorithm which trains each client for several local epochs and then the weight of each model is aggregated. Moreover we perform experiments across imbalanced and noisy data so as to explore scalability and robustness based on real-world datasets were noise is present and we also propose two innovative algorithms where the FL scenario is considered as a peer-to-peer formulation. Ultimately, to ensure that each device is not oversampled a client-balancing Dirichlet sampling algorithm with probabilistic guarantees is proposed.","PeriodicalId":68279,"journal":{"name":"计算机工程与设计","volume":"26 1","pages":"1-9"},"PeriodicalIF":0.0000,"publicationDate":"2022-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"计算机工程与设计","FirstCategoryId":"1093","ListUrlMain":"https://doi.org/10.1109/SEEDA-CECNSM57760.2022.9932980","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
Federated Learning (FL) is an emerging technique that assures user privacy and data integrity in distributed machine learning environments. To perform so, chunks of data are trained across edge devices and a high performance cluster server maintains a local copy without exchanging it with other parties. In this work, we investigate a FL scenario in a real-world case study using 5, 10 and 20 Raspberry Pi devices acting as clients. Under this setup, we employ the widely known FedAvg algorithm which trains each client for several local epochs and then the weight of each model is aggregated. Moreover we perform experiments across imbalanced and noisy data so as to explore scalability and robustness based on real-world datasets were noise is present and we also propose two innovative algorithms where the FL scenario is considered as a peer-to-peer formulation. Ultimately, to ensure that each device is not oversampled a client-balancing Dirichlet sampling algorithm with probabilistic guarantees is proposed.
期刊介绍:
Computer Engineering and Design is supervised by China Aerospace Science and Industry Corporation and sponsored by the 706th Institute of the Second Academy of China Aerospace Science and Industry Corporation. It was founded in 1980. The purpose of the journal is to disseminate new technologies and promote academic exchanges. Since its inception, it has adhered to the principle of combining depth and breadth, theory and application, and focused on reporting cutting-edge and hot computer technologies. The journal accepts academic papers with innovative and independent academic insights, including papers on fund projects, award-winning research papers, outstanding papers at academic conferences, doctoral and master's theses, etc.