{"title":"边缘设备网络中基于数据和模型并行的分布式深度学习系统","authors":"Tanmoy Sen, Haiying Shen","doi":"10.1109/ICCCN58024.2023.10230190","DOIUrl":null,"url":null,"abstract":"With the emergence of edge computing along with its local computation advantage over the cloud, methods for distributed deep learning (DL) training on edge nodes have been proposed. The increasing scale of DL models and large training dataset poses a challenge to run such jobs in one edge node due to resource constraints. However, the proposed methods either run the entire model in one edge node, collect all training data into one edge node, or still involve the remote cloud. To handle the challenge, we propose a fully distributed training system that realizes both Data and Model Parallelism over a network of edge devices (called DMP). It clusters the edge nodes to build a training structure by taking advantage of the feature that distributed edge nodes sense data for training. For each cluster, we propose a heuristic and a Reinforcement Learning (RL) based algorithm to handle the problem of how to partition a DL model and assign the partitions to edge nodes for model parallelism to minimize the overall training time. Taking advantage of the feature that geographically close edge nodes sense similar data, we further propose two schemes to avoid transferring duplicated data to the first-layer edge node as training data without compromising accuracy. Our container-based emulation and real edge node experiments show that our systems reduce up to 44% training time while maintaining the accuracy comparing with the state-of-the-art approaches. We also open sourced our source code.","PeriodicalId":132030,"journal":{"name":"2023 32nd International Conference on Computer Communications and Networks (ICCCN)","volume":"115 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"A Data and Model Parallelism based Distributed Deep Learning System in a Network of Edge Devices\",\"authors\":\"Tanmoy Sen, Haiying Shen\",\"doi\":\"10.1109/ICCCN58024.2023.10230190\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the emergence of edge computing along with its local computation advantage over the cloud, methods for distributed deep learning (DL) training on edge nodes have been proposed. The increasing scale of DL models and large training dataset poses a challenge to run such jobs in one edge node due to resource constraints. However, the proposed methods either run the entire model in one edge node, collect all training data into one edge node, or still involve the remote cloud. To handle the challenge, we propose a fully distributed training system that realizes both Data and Model Parallelism over a network of edge devices (called DMP). It clusters the edge nodes to build a training structure by taking advantage of the feature that distributed edge nodes sense data for training. For each cluster, we propose a heuristic and a Reinforcement Learning (RL) based algorithm to handle the problem of how to partition a DL model and assign the partitions to edge nodes for model parallelism to minimize the overall training time. Taking advantage of the feature that geographically close edge nodes sense similar data, we further propose two schemes to avoid transferring duplicated data to the first-layer edge node as training data without compromising accuracy. Our container-based emulation and real edge node experiments show that our systems reduce up to 44% training time while maintaining the accuracy comparing with the state-of-the-art approaches. We also open sourced our source code.\",\"PeriodicalId\":132030,\"journal\":{\"name\":\"2023 32nd International Conference on Computer Communications and Networks (ICCCN)\",\"volume\":\"115 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 32nd International Conference on Computer Communications and Networks (ICCCN)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCCN58024.2023.10230190\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 32nd International Conference on Computer Communications and Networks (ICCCN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCCN58024.2023.10230190","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Data and Model Parallelism based Distributed Deep Learning System in a Network of Edge Devices
With the emergence of edge computing along with its local computation advantage over the cloud, methods for distributed deep learning (DL) training on edge nodes have been proposed. The increasing scale of DL models and large training dataset poses a challenge to run such jobs in one edge node due to resource constraints. However, the proposed methods either run the entire model in one edge node, collect all training data into one edge node, or still involve the remote cloud. To handle the challenge, we propose a fully distributed training system that realizes both Data and Model Parallelism over a network of edge devices (called DMP). It clusters the edge nodes to build a training structure by taking advantage of the feature that distributed edge nodes sense data for training. For each cluster, we propose a heuristic and a Reinforcement Learning (RL) based algorithm to handle the problem of how to partition a DL model and assign the partitions to edge nodes for model parallelism to minimize the overall training time. Taking advantage of the feature that geographically close edge nodes sense similar data, we further propose two schemes to avoid transferring duplicated data to the first-layer edge node as training data without compromising accuracy. Our container-based emulation and real edge node experiments show that our systems reduce up to 44% training time while maintaining the accuracy comparing with the state-of-the-art approaches. We also open sourced our source code.