Bing Tang, Haiyan Li, Wei Xu, Buqing Cao, Qing Yang
{"title":"边缘计算中基于强化学习的服务放置能量延迟权衡","authors":"Bing Tang, Haiyan Li, Wei Xu, Buqing Cao, Qing Yang","doi":"10.1002/cpe.70154","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>Microservice technology, as a flexible application architecture, has gained wide popularity in the field of Internet of Things (IoT). IoT applications are highly sensitive to latency, making it crucial to place microservices on appropriate edge servers in an edge computing environment. Failure to do so can significantly impact service quality and degrade user experience, posing a major challenge. Addressing the aforementioned issues, this paper proposes a multiobjective service deployment strategy for IoT devices based on reinforcement learning. The goal is to minimize service access delay for IoT devices and reduce the average energy consumption of edge servers in the context of mobile edge computing. To achieve this, we first establish a stochastic optimization model using the Markov decision process (MDP) framework to handle service deployment and resource allocation dynamically. This model captures key characteristics such as heterogeneity in edge server capabilities, dynamic geographic information of IoT devices, and uncertainty in microservice requests. To overcome challenges related to dimensionality, slow convergence, and the exploration–exploitation tradeoff in traditional reinforcement learning algorithms, we introduce deep reinforcement learning into the optimization of microservice deployment. Specifically, we propose the use of deep deterministic policy gradient (DDPG) to obtain a near-optimal service deployment strategy without manual instructions. DDPG leverages the depth of the network to guide policy gradients and generate solutions that effectively balance exploration and exploitation. To evaluate the proposed approach, we implement the DPG-MSP (<span><b>DDPG</b></span>-based <span><b>M</b></span>icro<span><b>S</b></span>ervice <span><b>P</b></span>lacement) algorithm using real datasets and synthetic data. Comparative analysis with existing microservice deployment algorithms demonstrates the superiority of DDPG-MSP in terms of performance, robustness, and scalability.</p>\n </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"37 15-17","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2025-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Energy-Latency Tradeoffs for Service Placement Based on Reinforcement Learning in Edge Computing\",\"authors\":\"Bing Tang, Haiyan Li, Wei Xu, Buqing Cao, Qing Yang\",\"doi\":\"10.1002/cpe.70154\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n <p>Microservice technology, as a flexible application architecture, has gained wide popularity in the field of Internet of Things (IoT). IoT applications are highly sensitive to latency, making it crucial to place microservices on appropriate edge servers in an edge computing environment. Failure to do so can significantly impact service quality and degrade user experience, posing a major challenge. Addressing the aforementioned issues, this paper proposes a multiobjective service deployment strategy for IoT devices based on reinforcement learning. The goal is to minimize service access delay for IoT devices and reduce the average energy consumption of edge servers in the context of mobile edge computing. To achieve this, we first establish a stochastic optimization model using the Markov decision process (MDP) framework to handle service deployment and resource allocation dynamically. This model captures key characteristics such as heterogeneity in edge server capabilities, dynamic geographic information of IoT devices, and uncertainty in microservice requests. To overcome challenges related to dimensionality, slow convergence, and the exploration–exploitation tradeoff in traditional reinforcement learning algorithms, we introduce deep reinforcement learning into the optimization of microservice deployment. Specifically, we propose the use of deep deterministic policy gradient (DDPG) to obtain a near-optimal service deployment strategy without manual instructions. DDPG leverages the depth of the network to guide policy gradients and generate solutions that effectively balance exploration and exploitation. To evaluate the proposed approach, we implement the DPG-MSP (<span><b>DDPG</b></span>-based <span><b>M</b></span>icro<span><b>S</b></span>ervice <span><b>P</b></span>lacement) algorithm using real datasets and synthetic data. Comparative analysis with existing microservice deployment algorithms demonstrates the superiority of DDPG-MSP in terms of performance, robustness, and scalability.</p>\\n </div>\",\"PeriodicalId\":55214,\"journal\":{\"name\":\"Concurrency and Computation-Practice & Experience\",\"volume\":\"37 15-17\",\"pages\":\"\"},\"PeriodicalIF\":1.5000,\"publicationDate\":\"2025-06-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Concurrency and Computation-Practice & Experience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/cpe.70154\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, SOFTWARE ENGINEERING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Concurrency and Computation-Practice & Experience","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/cpe.70154","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
Energy-Latency Tradeoffs for Service Placement Based on Reinforcement Learning in Edge Computing
Microservice technology, as a flexible application architecture, has gained wide popularity in the field of Internet of Things (IoT). IoT applications are highly sensitive to latency, making it crucial to place microservices on appropriate edge servers in an edge computing environment. Failure to do so can significantly impact service quality and degrade user experience, posing a major challenge. Addressing the aforementioned issues, this paper proposes a multiobjective service deployment strategy for IoT devices based on reinforcement learning. The goal is to minimize service access delay for IoT devices and reduce the average energy consumption of edge servers in the context of mobile edge computing. To achieve this, we first establish a stochastic optimization model using the Markov decision process (MDP) framework to handle service deployment and resource allocation dynamically. This model captures key characteristics such as heterogeneity in edge server capabilities, dynamic geographic information of IoT devices, and uncertainty in microservice requests. To overcome challenges related to dimensionality, slow convergence, and the exploration–exploitation tradeoff in traditional reinforcement learning algorithms, we introduce deep reinforcement learning into the optimization of microservice deployment. Specifically, we propose the use of deep deterministic policy gradient (DDPG) to obtain a near-optimal service deployment strategy without manual instructions. DDPG leverages the depth of the network to guide policy gradients and generate solutions that effectively balance exploration and exploitation. To evaluate the proposed approach, we implement the DPG-MSP (DDPG-based MicroService Placement) algorithm using real datasets and synthetic data. Comparative analysis with existing microservice deployment algorithms demonstrates the superiority of DDPG-MSP in terms of performance, robustness, and scalability.
期刊介绍:
Concurrency and Computation: Practice and Experience (CCPE) publishes high-quality, original research papers, and authoritative research review papers, in the overlapping fields of:
Parallel and distributed computing;
High-performance computing;
Computational and data science;
Artificial intelligence and machine learning;
Big data applications, algorithms, and systems;
Network science;
Ontologies and semantics;
Security and privacy;
Cloud/edge/fog computing;
Green computing; and
Quantum computing.