Deep Smart Scheduling: A Deep Learning Approach for Automated Big Data Scheduling Over the Cloud

Gaith Rjoub, J. Bentahar, O. A. Wahab, A. Bataineh
{"title":"Deep Smart Scheduling: A Deep Learning Approach for Automated Big Data Scheduling Over the Cloud","authors":"Gaith Rjoub, J. Bentahar, O. A. Wahab, A. Bataineh","doi":"10.1109/FiCloud.2019.00034","DOIUrl":null,"url":null,"abstract":"With the widespread adoption of Internet of Thing (IoT) and the exponential growth in the volumes of generated data, cloud providers tend to receive massive waves of demands on their storage and computing resources. To help providers deal with such demands without sacrificing performance, the concept of cloud automation had recently arisen to improve the performance and reduce the manual efforts related to the management of cloud computing workloads. In this context, we propose in this paper, Deep learning Smart Scheduling (DSS), an automated big data task scheduling approach in cloud computing environments. DSS combines Deep Reinforcement Learning (DRL) and Long Short-Term Memory (LSTM) to automatically predict the Virtual Machines (VMs) to which each incoming big data task should be scheduled to so as to improve the performance of big data analytics and reduce their resource execution cost. Experiments conducted using real-world datasets from Google Cloud Platform show that our solution minimizes the CPU usage cost by 28.8% compared to the Shortest Job First (SJF), and by 14% compared to both the Round Robin (RR) and improved Particle Swarm Optimization (PSO) approaches. Moreover, our solution decreases the RAM memory usage cost by 31.25% compared to the SJF, by 25% compared to the RR, and by 18.78% compared to the improved PSO.","PeriodicalId":268882,"journal":{"name":"2019 7th International Conference on Future Internet of Things and Cloud (FiCloud)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"26","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 7th International Conference on Future Internet of Things and Cloud (FiCloud)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/FiCloud.2019.00034","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 26

Abstract

With the widespread adoption of Internet of Thing (IoT) and the exponential growth in the volumes of generated data, cloud providers tend to receive massive waves of demands on their storage and computing resources. To help providers deal with such demands without sacrificing performance, the concept of cloud automation had recently arisen to improve the performance and reduce the manual efforts related to the management of cloud computing workloads. In this context, we propose in this paper, Deep learning Smart Scheduling (DSS), an automated big data task scheduling approach in cloud computing environments. DSS combines Deep Reinforcement Learning (DRL) and Long Short-Term Memory (LSTM) to automatically predict the Virtual Machines (VMs) to which each incoming big data task should be scheduled to so as to improve the performance of big data analytics and reduce their resource execution cost. Experiments conducted using real-world datasets from Google Cloud Platform show that our solution minimizes the CPU usage cost by 28.8% compared to the Shortest Job First (SJF), and by 14% compared to both the Round Robin (RR) and improved Particle Swarm Optimization (PSO) approaches. Moreover, our solution decreases the RAM memory usage cost by 31.25% compared to the SJF, by 25% compared to the RR, and by 18.78% compared to the improved PSO.
深度智能调度:基于云的自动化大数据调度的深度学习方法
随着物联网(IoT)的广泛采用和生成数据量的指数级增长,云提供商往往会收到对其存储和计算资源的大量需求。为了帮助提供商在不牺牲性能的情况下处理这些需求,最近出现了云自动化的概念,以提高性能并减少与云计算工作负载管理相关的人工工作。在此背景下,我们在本文中提出了深度学习智能调度(DSS),一种云计算环境下的自动化大数据任务调度方法。DSS将DRL (Deep Reinforcement Learning)和LSTM (Long - Short-Term Memory)相结合,自动预测每个传入的大数据任务应该调度到哪些虚拟机上,从而提高大数据分析的性能,降低大数据分析的资源执行成本。使用来自Google Cloud Platform的真实数据集进行的实验表明,与最短作业优先(SJF)方法相比,我们的解决方案将CPU使用成本降低了28.8%,与轮询(RR)和改进粒子群优化(PSO)方法相比,降低了14%。此外,与SJF相比,我们的解决方案将RAM内存使用成本降低了31.25%,与RR相比降低了25%,与改进的PSO相比降低了18.78%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信