Task Assignment in a Virtualized GPU Enabled Cloud

Hari Sivaraman, Uday Kurkure, Lan Vu
{"title":"Task Assignment in a Virtualized GPU Enabled Cloud","authors":"Hari Sivaraman, Uday Kurkure, Lan Vu","doi":"10.1109/HPCS.2018.00143","DOIUrl":null,"url":null,"abstract":"Cloud computing vendors are beginning to offer GPU based high performance computing as a service. One approach uses virtual machines (VM), running in a hypervisor like VMware vSphere, equipped with virtual GPUs like Nvidia's vGPU solution. In this approach, multiple VMs running concurrently can share a single GPU. The number of VMs that share the GPU can be configured by the user/system administrator. Further, VMs can be re-assigned to GPUs, if more than one is available, dynamically. This approach allows tasks/jobs that use GPUs to run in individual VMs guaranteeing isolation whilst sharing resources. In a typical cloud environment with multiple servers each with one or more GPUs, finding an efficient, fast solution to the problem of placing VMs (i.e. VM-placement) on GPUs and moving them around as needed is extremely important to achieve high throughput of tasks while maximizing server utilization and minimizing task wait times. In this paper, we present the simulator we built to compare different solutions to the problem of VM-placement together with some early results.","PeriodicalId":308138,"journal":{"name":"2018 International Conference on High Performance Computing & Simulation (HPCS)","volume":"98 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 International Conference on High Performance Computing & Simulation (HPCS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/HPCS.2018.00143","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

Cloud computing vendors are beginning to offer GPU based high performance computing as a service. One approach uses virtual machines (VM), running in a hypervisor like VMware vSphere, equipped with virtual GPUs like Nvidia's vGPU solution. In this approach, multiple VMs running concurrently can share a single GPU. The number of VMs that share the GPU can be configured by the user/system administrator. Further, VMs can be re-assigned to GPUs, if more than one is available, dynamically. This approach allows tasks/jobs that use GPUs to run in individual VMs guaranteeing isolation whilst sharing resources. In a typical cloud environment with multiple servers each with one or more GPUs, finding an efficient, fast solution to the problem of placing VMs (i.e. VM-placement) on GPUs and moving them around as needed is extremely important to achieve high throughput of tasks while maximizing server utilization and minimizing task wait times. In this paper, we present the simulator we built to compare different solutions to the problem of VM-placement together with some early results.
虚拟化GPU云中的任务分配
云计算供应商开始提供基于GPU的高性能计算服务。一种方法是使用虚拟机(VM),运行在像VMware vSphere这样的管理程序中,配备像Nvidia的vGPU解决方案这样的虚拟gpu。多个并发运行的虚拟机可以共享一个GPU。共享图形处理器的虚拟机数量可由用户/系统管理员配置。此外,如果有多个可用gpu,可以动态地将虚拟机重新分配给gpu。这种方法允许使用gpu的任务/作业在单独的vm中运行,在共享资源的同时保证隔离。在典型的云环境中,有多个服务器,每个服务器都有一个或多个gpu,找到一个高效、快速的解决方案来解决在gpu上放置vm(即vm放置)并根据需要移动它们的问题,这对于实现任务的高吞吐量,同时最大化服务器利用率和最小化任务等待时间非常重要。在本文中,我们展示了我们构建的模拟器,以比较vm放置问题的不同解决方案以及一些早期结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信