{"title":"Maximizing profitability through cloud-enabled Reinforcement Learning for UAV coverage in real-time e-business applications","authors":"Haythem Bany Salameh , Ghaleb Elrefae , Mohannad Alhafnawi , Yaser Jararweh , Ayat Alkhdour , Sharief Abdel-Razeq","doi":"10.1016/j.simpat.2024.102970","DOIUrl":null,"url":null,"abstract":"<div><p>Real-time e-business applications are vital for operational efficiency, but connectivity challenges persist, particularly in remote or crowded areas. Drone Base Station (DBS) architecture, proposed for Beyond fifth Generation (B5G) and Sixth Generation (6G) multi-cell networks, offers on-demand hotspot coverage, addressing connectivity gaps in remote or crowded environments. DBSs provide a promising solution to meet the demanding requirements of high data rates, real-time responsiveness, low latency, and extended network coverage, particularly for real-time e-business applications. A critical challenge in this context involves efficiently allocating the needed number of DBSs to the different hotspot service areas, referred to as cells, to optimize the operator’s total profit under unpredictable user demands, varying area-specific service costs, and price dependence real-time e-service. The objective is to achieve the highest total revenue while minimizing the cost (cost savings) throughout the multi-cell system. This challenge is formulated as a profit-maximization discount return problem, integrating the coverage constraint, the variable cell-dependent operational cost, the e-service-based price and the uncertain demands of users across cells. Traditional optimization methods fail due to environmental uncertainty, which leads to the need to reformulate the problem as a Markov Decision Problem (MDP). We introduce a cloud-based Reinforcement Learning (RL) algorithm for DBS dispatch to address the MDP formulation. This algorithm dynamically adjusts to uncertain per-cell user distributions, considering variable operating costs and service-dependent prices across cells. Through extensive evaluation, the RL-based dispatch approach is compared with reference drone dispatch algorithms, demonstrating superior performance in maximizing operator profit through <em>cost savings</em> by optimizing DBS dispatch decisions based on learned user behaviors, variable operational costs, and e-service types.</p></div>","PeriodicalId":49518,"journal":{"name":"Simulation Modelling Practice and Theory","volume":"135 ","pages":"Article 102970"},"PeriodicalIF":3.5000,"publicationDate":"2024-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Simulation Modelling Practice and Theory","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1569190X24000844","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0
Abstract
Real-time e-business applications are vital for operational efficiency, but connectivity challenges persist, particularly in remote or crowded areas. Drone Base Station (DBS) architecture, proposed for Beyond fifth Generation (B5G) and Sixth Generation (6G) multi-cell networks, offers on-demand hotspot coverage, addressing connectivity gaps in remote or crowded environments. DBSs provide a promising solution to meet the demanding requirements of high data rates, real-time responsiveness, low latency, and extended network coverage, particularly for real-time e-business applications. A critical challenge in this context involves efficiently allocating the needed number of DBSs to the different hotspot service areas, referred to as cells, to optimize the operator’s total profit under unpredictable user demands, varying area-specific service costs, and price dependence real-time e-service. The objective is to achieve the highest total revenue while minimizing the cost (cost savings) throughout the multi-cell system. This challenge is formulated as a profit-maximization discount return problem, integrating the coverage constraint, the variable cell-dependent operational cost, the e-service-based price and the uncertain demands of users across cells. Traditional optimization methods fail due to environmental uncertainty, which leads to the need to reformulate the problem as a Markov Decision Problem (MDP). We introduce a cloud-based Reinforcement Learning (RL) algorithm for DBS dispatch to address the MDP formulation. This algorithm dynamically adjusts to uncertain per-cell user distributions, considering variable operating costs and service-dependent prices across cells. Through extensive evaluation, the RL-based dispatch approach is compared with reference drone dispatch algorithms, demonstrating superior performance in maximizing operator profit through cost savings by optimizing DBS dispatch decisions based on learned user behaviors, variable operational costs, and e-service types.
期刊介绍:
The journal Simulation Modelling Practice and Theory provides a forum for original, high-quality papers dealing with any aspect of systems simulation and modelling.
The journal aims at being a reference and a powerful tool to all those professionally active and/or interested in the methods and applications of simulation. Submitted papers will be peer reviewed and must significantly contribute to modelling and simulation in general or use modelling and simulation in application areas.
Paper submission is solicited on:
• theoretical aspects of modelling and simulation including formal modelling, model-checking, random number generators, sensitivity analysis, variance reduction techniques, experimental design, meta-modelling, methods and algorithms for validation and verification, selection and comparison procedures etc.;
• methodology and application of modelling and simulation in any area, including computer systems, networks, real-time and embedded systems, mobile and intelligent agents, manufacturing and transportation systems, management, engineering, biomedical engineering, economics, ecology and environment, education, transaction handling, etc.;
• simulation languages and environments including those, specific to distributed computing, grid computing, high performance computers or computer networks, etc.;
• distributed and real-time simulation, simulation interoperability;
• tools for high performance computing simulation, including dedicated architectures and parallel computing.