Shujuan Tian;Shuhuan Xiang;Ziqi Zhou;Haipeng Dai;Enze Yu;Qingyong Deng
{"title":"Task Offloading and Resource Allocation Based on Reinforcement Learning and Load Balancing in Vehicular Networking","authors":"Shujuan Tian;Shuhuan Xiang;Ziqi Zhou;Haipeng Dai;Enze Yu;Qingyong Deng","doi":"10.1109/TCE.2025.3542133","DOIUrl":null,"url":null,"abstract":"Due to limited on-board resources and the mobility characteristics of vehicles in a multi-access edge computing (MEC)-based vehicular network, efficient task offloading and resource allocation schemes are essential for achieving low-latency and low-energy consumption applications in the Internet of Vehicles (IoV). The spatial distribution of vehicles, influenced by various factors, leads to significant workload variations across MEC servers. In this paper, we address task offloading and resource allocation as a joint optimization problem and propose a Load-Balancing Deep Deterministic Policy Gradient (LBDDPG) algorithm to achieve optimal results. The joint optimization problem is modeled as a Markov Decision Process (MDP), enabling the LBDDPG algorithm to systematically address the challenges of workload imbalance and resource inefficiency. The algorithm incorporates a load optimization strategy to balance workload distribution across MEC servers, mitigating disparities caused by uneven vehicle distributions. The reward function is designed to account for both energy consumption and delay, ensuring an optimal trade-off between these critical factors. To enhance training efficiency, a noise-based exploration strategy is employed, preventing ineffective exploration during the early stages. Additionally, constraints such as computational capacity and latency thresholds are embedded to ensure the algorithm’s practical applicability. Experimental results demonstrate that the proposed LBDDPG algorithm achieves faster convergence and superior performance in terms of energy consumption and latency compared to other reinforcement learning algorithms.","PeriodicalId":13208,"journal":{"name":"IEEE Transactions on Consumer Electronics","volume":"71 1","pages":"2217-2230"},"PeriodicalIF":4.3000,"publicationDate":"2025-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Consumer Electronics","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10887320/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Due to limited on-board resources and the mobility characteristics of vehicles in a multi-access edge computing (MEC)-based vehicular network, efficient task offloading and resource allocation schemes are essential for achieving low-latency and low-energy consumption applications in the Internet of Vehicles (IoV). The spatial distribution of vehicles, influenced by various factors, leads to significant workload variations across MEC servers. In this paper, we address task offloading and resource allocation as a joint optimization problem and propose a Load-Balancing Deep Deterministic Policy Gradient (LBDDPG) algorithm to achieve optimal results. The joint optimization problem is modeled as a Markov Decision Process (MDP), enabling the LBDDPG algorithm to systematically address the challenges of workload imbalance and resource inefficiency. The algorithm incorporates a load optimization strategy to balance workload distribution across MEC servers, mitigating disparities caused by uneven vehicle distributions. The reward function is designed to account for both energy consumption and delay, ensuring an optimal trade-off between these critical factors. To enhance training efficiency, a noise-based exploration strategy is employed, preventing ineffective exploration during the early stages. Additionally, constraints such as computational capacity and latency thresholds are embedded to ensure the algorithm’s practical applicability. Experimental results demonstrate that the proposed LBDDPG algorithm achieves faster convergence and superior performance in terms of energy consumption and latency compared to other reinforcement learning algorithms.
期刊介绍:
The main focus for the IEEE Transactions on Consumer Electronics is the engineering and research aspects of the theory, design, construction, manufacture or end use of mass market electronics, systems, software and services for consumers.