{"title":"Joint Autoscaling of Containers and Virtual Machines for Cost Optimization in Container Clusters","authors":"","doi":"10.1007/s10723-023-09732-4","DOIUrl":"https://doi.org/10.1007/s10723-023-09732-4","url":null,"abstract":"<h3>Abstract</h3> <p>Autoscaling enables container cluster orchestrators to automatically adjust computational resources, such as containers and Virtual Machines (VMs), to handle fluctuating workloads effectively. This adaptation can involve modifying the amount of resources (horizontal scaling) or adjusting their computational capacity (vertical scaling). The motivation for our work stems from the limitations of previous autoscaling approaches, which are either partial (scaling containers or VMs, but not both) or excessively complex to be used in real systems. This complexity arises from their use of models with a large number of variables and the addressing of two simultaneous challenges: achieving the optimal deployment for a single scheduling window and managing the transition between successive scheduling windows. We propose an Integer Linear Programming (ILP) model to address the challenge of autoscaling containers and VMs jointly, both horizontally and vertically, to minimize deployment costs. This model is designed to be used with predictive autoscalers and be solved in a reasonable time, even for large clusters. To this end, improvements and reasonable simplifications with respect to previous models have been carried out to drastically reduce the size of the resource allocation problem. Furthermore, the proposed model provides an enhanced representation of system performance in comparison to previous approaches. A tool called Conlloovia has been developed to implement this model. To evaluate its performance, we have conducted a comprehensive assessment, comparing it with two heuristic allocators with different problem sizes. Our findings indicate that Conlloovia consistently demonstrates lower deployment costs in a significant number of cases. Conlloovia has also been evaluated with a real application, using synthetic and real workload traces, as well as different scheduling windows, with deployment costs approximately 20% lower than heuristic allocators.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"44 1","pages":""},"PeriodicalIF":5.5,"publicationDate":"2024-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139562459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Intrusion Detection using Federated Attention Neural Network for Edge Enabled Internet of Things","authors":"Xiedong Song, Qinmin Ma","doi":"10.1007/s10723-023-09725-3","DOIUrl":"https://doi.org/10.1007/s10723-023-09725-3","url":null,"abstract":"<p>Edge nodes, which are expected to grow into a multi-billion-dollar market, are essential for detection against a variety of cyber threats on Internet-of-Things endpoints. Adopting the current network intrusion detection system with deep learning models (DLM) based on FedACNN is constrained by the resource limitations of this network equipment layer. We solve this issue by creating a unique, lightweight, quick, and accurate edge detection model to identify DLM-based distributed denial service attacks on edge nodes. Our approach can generate real results at a relevant pace even with limited resources, such as low power, memory, and processing capabilities. The Federated Convolution Neural Network (FedACNN) deep learning method uses attention mechanisms to minimise communication delay. The developed model uses a recent cybersecurity dataset deployed on an edge node simulated by a Raspberry Pi (UNSW 2015). Our findings show that, compared to traditional DLM methodologies, our model retains a high accuracy rate of about 99%, even with decreased CPU and memory resource use. Also, it is about three times smaller in volume than the most advanced model while requiring a lot less testing time.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"60 1","pages":""},"PeriodicalIF":5.5,"publicationDate":"2024-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139509469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Y. P. Tsang, C. K. M. Lee, Kening Zhang, C. H. Wu, W. H. Ip
{"title":"On-Chain and Off-Chain Data Management for Blockchain-Internet of Things: A Multi-Agent Deep Reinforcement Learning Approach","authors":"Y. P. Tsang, C. K. M. Lee, Kening Zhang, C. H. Wu, W. H. Ip","doi":"10.1007/s10723-023-09739-x","DOIUrl":"https://doi.org/10.1007/s10723-023-09739-x","url":null,"abstract":"<p>The emergence of blockchain technology has seen applications increasingly hybridise cloud storage and distributed ledger technology in the Internet of Things (IoT) and cyber-physical systems, complicating data management in decentralised applications (DApps). Because it is inefficient for blockchain technology to handle large amounts of data, effective on-chain and off-chain data management in peer-to-peer networks and cloud storage has drawn considerable attention. Space reservation is a cost-effective approach to managing cloud storage effectively, contrasting with the demand for additional space in real-time. Furthermore, off-chain data replication in the peer-to-peer network can eliminate single points of failure of DApps. However, recent research has rarely discussed optimising on-chain and off-chain data management in the blockchain-enabled IoT (BIoT) environment. In this study, the BIoT environment is modelled, with cloud storage and blockchain orchestrated over the peer-to-peer network. The asynchronous advantage actor-critic algorithm is applied to exploit intelligent agents with the optimal policy for data packing, space reservation, and data replication to achieve an intelligent data management strategy. The experimental analysis reveals that the proposed scheme demonstrates rapid convergence and superior performance in terms of average total reward compared with other typical schemes, resulting in enhanced scalability, security and reliability of blockchain-IoT networks, leading to an intelligent data management strategy.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"32 1","pages":""},"PeriodicalIF":5.5,"publicationDate":"2024-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139509263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"3D Lidar Target Detection Method at the Edge for the Cloud Continuum","authors":"Xuemei Li, Xuelian Liu, Da Xie, Chong Chen","doi":"10.1007/s10723-023-09736-0","DOIUrl":"https://doi.org/10.1007/s10723-023-09736-0","url":null,"abstract":"<p>In the internet of things, machine learning at the edge of cloud continuum is developing rapidly, providing more convenient services for design developers. The paper proposes a lidar target detection method based on scene density-awareness network for cloud continuum. The density-awareness network architecture is designed, and the context column feature network is proposed. The BEV density attention feature network is designed by cascading the density feature map with the spatial attention mechanism, and then connected with the BEV column feature network to generate the ablation BEV map. Multi-head detector is designed to regress the object center point, scale size and direction, and loss function is used for active supervision. The experiment is conducted on Alibaba Cloud services. On the validation dataset of KITTI, the 3D objects and BEV objects are detected and evaluated for three types of objects. The results show that most of the AP values of the density-awareness model proposed in this paper are higher than other methods, and the detection time is 0.09 s, which can meet the requirements of high accuracy and real-time of vehicle-borne lidar target detection.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"112 1","pages":""},"PeriodicalIF":5.5,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139509072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"AI-Driven Task Scheduling Strategy with Blockchain Integration for Edge Computing","authors":"Avishek Sinha, Samayveer Singh, Harsh K. Verma","doi":"10.1007/s10723-024-09743-9","DOIUrl":"https://doi.org/10.1007/s10723-024-09743-9","url":null,"abstract":"<p>In recent times, edge computing has arisen as a highly promising paradigm aimed at facilitating resource-intensive Internet of Things (IoT) applications by offering low-latency services. However, the constrained computational capabilities of the IoT nodes present considerable obstacles when it comes to efficient task-scheduling applications. In this paper, a nature-inspired coati optimization-based energy-aware task scheduling (CO-ETS) approach is proposed to address the challenge of efficiently assigning tasks to available edge devices. The proposed work incorporates a fitness function that effectively enhances task assignment optimization, leading to improved system efficiency, reduced power consumption, and enhanced system reliability. Moreover, we integrate blockchain with AI-driven task scheduling to fortify security, protect user privacy, and optimize edge computing in IoT-based environments. The blockchain-based approach ensures a secure and trusted decentralized identity management and reputation system for IoT edge networks. To validate the effectiveness of the proposed CO-ETS approach, we conduct a comparative analysis against state-of-the-art methods by considering metrics such as makespan, CPU execution time, energy consumption, and mean wait time. The proposed approach offers promising solutions to optimize task allocation, enhance system performance, and ensure secure and privacy-preserving operations in edge computing environments.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"29 1","pages":""},"PeriodicalIF":5.5,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139509189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Optimizing Accounting Informatization through Simultaneous Multi-Tasking across Edge and Cloud Devices using Hybrid Machine Learning Models","authors":"Xiaofeng Yang","doi":"10.1007/s10723-023-09735-1","DOIUrl":"https://doi.org/10.1007/s10723-023-09735-1","url":null,"abstract":"<p>Accounting informatization is a crucial component of enterprise informatization, significantly impacting operational efficiency in accounting and finance. Advances in information technology have introduced automation techniques that accelerate the processing of accounting information cost-effectively. Integrating artificial intelligence, cloud computing, and edge computing is pivotal in streamlining and optimizing these processes. Traditionally, accounting informatization relied on system servers and local storage for data processing. However, the era of big data necessitates a shift to cloud computing frameworks for efficient data storage and processing. Despite the advantages of cloud storage, concerns arise regarding data security and the substantial data transactions between the cloud and source devices. To address these challenges, this research proposes a novel algorithm, Heterogeneous Distributed Deep Learning with Data Offloading (DDLO) algorithm. DDLO leverages the synergy between edge devices and cloud computing to enhance data processes. Edge computing enables rapid processing of large volumes of data at or near the data collection sites, optimizing day-to-day operations for enterprises. Furthermore, machine learning algorithms at edge devices enhance data processing efficiency, augmenting the computing environment's overall performance. The proposed DDLO algorithm fosters a hybrid machine learning approach for computing joint tasks and multi-tasking in accounting informatization. It enables dynamic resource allocation, allowing selected data or model updates to be offloaded to the cloud for complex tasks. The algorithm's performance is rigorously evaluated using key metrics, including computing time, offloading time, accuracy, and cost levels. By capitalizing on the strengths of edge computing, cloud computing, and artificial intelligence, the DDLO algorithm effectively addresses accounting informatization challenges. It empowers enterprises to process vast amounts of accounting data efficiently and securely while improving overall operational efficiency. Regarding time, using terasort in tasks offloading using DDLO consumes less milliseconds 0t 33 ms which is lesser than other techniques.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"9 1","pages":""},"PeriodicalIF":5.5,"publicationDate":"2024-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139501489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"CMSV: a New Cloud Multi-Agents for Self-Driving Vehicles as a Services","authors":"Aida A. Nasr","doi":"10.1007/s10723-023-09734-2","DOIUrl":"https://doi.org/10.1007/s10723-023-09734-2","url":null,"abstract":"","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"24 3","pages":"1-13"},"PeriodicalIF":5.5,"publicationDate":"2024-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139437563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hybrid Fuzzy Neural Network for Joint Task Offloading in the Internet of Vehicles","authors":"Bingtao Liu","doi":"10.1007/s10723-023-09724-4","DOIUrl":"https://doi.org/10.1007/s10723-023-09724-4","url":null,"abstract":"<p>The Internet of Vehicles (IoV) technology is progressively maturing because of the growth of private cars and the establishment of intelligent transportation systems. The development of smart cars has, therefore, been followed by a parallel rise in the volume of media and video games in the automobile and a massive increase in the need for processing resources. Smart cars cannot process the enormous quantity of requests created by vehicles because they have limited computing power and must maintain many outstanding jobs in their queues. The distribution of edge servers near the customer side of the highway may also accomplish real-time resource requests, and edge servers can assist with the shortage of computational power. Nevertheless, the substantial amount of energy created while processing is also an issue we must address. A joint task offloading strategy based on mobile edge computing and fog computing (EFTO) was presented in this paper to address this problem. Practically, the position of the processing activity is first discovered by obtaining the computing task's route, which reveals all the task's routing details from the starting point to the desired place. Next, to minimize the time and time expended during offloading and processing, a multi-objective optimization problem is implemented using the task offloading technique F-TORA based on the Takagi–Sugeno fuzzy neural network (T-S FNN). Finally, comparative trials showing a decrease in time consumed and an optimization of energy use compared to alternative offloading techniques prove the effectiveness of EFTO.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"80 3 1","pages":""},"PeriodicalIF":5.5,"publicationDate":"2024-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139414716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Parking Cooperation-Based Mobile Edge Computing Using Task Offloading Strategy","authors":"XuanWen, Hai Meng Sun","doi":"10.1007/s10723-023-09721-7","DOIUrl":"https://doi.org/10.1007/s10723-023-09721-7","url":null,"abstract":"<p>The surge in computing demands of onboard devices in vehicles has necessitated the adoption of mobile edge computing (MEC) to cater to their computational and storage needs. This paper presents a task offloading strategy for mobile edge computing based on collaborative roadside parking cooperation, leveraging idle computing resources in roadside vehicles. The proposed method establishes resource sharing and mutual utilization among roadside vehicles, roadside units (RSUs), and cloud servers, transforming the computing task offloading problem into a constrained optimization challenge. To address the complexity of this optimization problem, a novel Hybrid Algorithm based on the Hill-Climbing and Genetic Algorithm (HHGA) is proposed, combined with the powerful Simulated Annealing (SA) algorithm. The HHGA-SA Algorithm integrates the advantages of both HHGA and SA to efficiently explore the solution space and optimize task execution with reduced delay and energy consumption. The HHGA component of the algorithm utilizes the strengths of Genetic Algorithm and Hill-Climbing. The Genetic Algorithm enables global exploration, identifying potential optimal solutions, while Hill-Climbing refines the solutions locally to improve their quality. By harnessing the synergies between these techniques, the HHGA-SA Algorithm navigates the multi-constraint landscape effectively, producing robust and high-quality solutions for task offloading. To evaluate the efficacy of the proposed approach, extensive simulations are conducted in a realistic roadside parking cooperation-based Mobile Edge Computing scenario. Comparative analyses with standard Genetic Algorithms and Hill-Climbing demonstrate the superiority of the HHGA-SA Algorithm, showcasing substantial enhancements in task execution efficiency and energy utilization. The study highlights the potential of leveraging idle computing resources in roadside parking vehicles to enhance Mobile Edge Computing capabilities. The collaborative approach facilitated by the HHGA-SA Algorithm fosters efficient task offloading among roadside vehicles, RSUs, and cloud servers, elevating overall system performance.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"140 1","pages":""},"PeriodicalIF":5.5,"publicationDate":"2024-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139398597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Cloud-Edge-Based Multi-Objective Task Scheduling Approach for Smart Manufacturing Lines","authors":"Huayi Yin, Xindong Huang, Erzhong Cao","doi":"10.1007/s10723-023-09723-5","DOIUrl":"https://doi.org/10.1007/s10723-023-09723-5","url":null,"abstract":"<p>The number of task demands created by smart terminals is rising dramatically because of the increasing usage of industrial Internet technologies in intelligent production lines. Speed of response is vital when dealing with such large activities. The current work needs to work with the task scheduling flow of smart manufacturing lines. The proposed method addresses the limitations of the current approach, particularly in the context of task scheduling and task scheduling flow within intelligent production lines. This study concentrates on solving the multi-objective task scheduling challenge in intelligent manufacturing by introducing a task scheduling approach based on job prioritization. To achieve this, a multi-objective task scheduling mechanism was developed, aiming to reduce service latency and energy consumption. This mechanism was integrated into a cloud-edge computing framework for intelligent production lines. The task scheduling strategy and task flow scheduling were optimized using Particle Swarm Optimization (PSO) and Gravitational Search Algorithm (GSA). Lastly, thorough simulation studies evaluate Multi-PSG, demonstrating that it beats every other algorithm regarding job completion rate. The completion rate of all tasks is greater than 90% when the number of nodes exceeds 10, which satisfies the real-time demands of the related tasks in the smart manufacturing processes. The method also performs better than other methods regarding power usage and maximum completion rate.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":"70 1","pages":""},"PeriodicalIF":5.5,"publicationDate":"2024-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139398585","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}