Cooling-aware and thermal-aware workload placement for green HPC data centers

Ayan Banerjee, T. Mukherjee, G. Varsamopoulos, S. Gupta
{"title":"Cooling-aware and thermal-aware workload placement for green HPC data centers","authors":"Ayan Banerjee, T. Mukherjee, G. Varsamopoulos, S. Gupta","doi":"10.1109/GREENCOMP.2010.5598306","DOIUrl":null,"url":null,"abstract":"High Performance Computing (HPC) data centers are becoming increasingly dense; the associated power-density and energy consumption of their operation is increasing. Up to half of the total energy is attributed to cooling the data center; greening the data center operations to reduce both computing and cooling energy is imperative. To this effect: i) the Energy Inefficiency Ratio of SPatial job scheduling (a.k.a. job placement) algorithms, also referred as SP-EIR, is analyzed by comparing the total (computing + cooling) energy consumption incurred by the algorithms with the minimum possible energy consumption, while assuming that the job start times are already decided to meet the Service Level Agreements (SLAs); and ii) a coordinated cooling-aware job placement and cooling management algorithm, Highest Thermostat Setting (HTS), is developed. HTS is aware of dynamic behavior of the Computer Room Air Conditioner (CRAC) units and places the jobs in a way to reduce the cooling demands from the CRACs. Dynamic updates of the CRAC thermostat settings based on the cooling demands can enable a reduction in energy consumption. Simulation results based on power measurements and job traces from the ASU HPC data center show that: i) HTS reduces the SP-EIR by 15% compared to LRH, a thermal-aware spatial scheduling algorithm; and ii) in conjunction with FCFS-Backfill, HTS increases the throughput per unit energy by 6.89% and 5.56%, respectively, over LRH and MTDP (an energy-effcient spatial scheduling algorithm with server consolidation).","PeriodicalId":262148,"journal":{"name":"International Conference on Green Computing","volume":"7 5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"94","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Green Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/GREENCOMP.2010.5598306","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 94

Abstract

High Performance Computing (HPC) data centers are becoming increasingly dense; the associated power-density and energy consumption of their operation is increasing. Up to half of the total energy is attributed to cooling the data center; greening the data center operations to reduce both computing and cooling energy is imperative. To this effect: i) the Energy Inefficiency Ratio of SPatial job scheduling (a.k.a. job placement) algorithms, also referred as SP-EIR, is analyzed by comparing the total (computing + cooling) energy consumption incurred by the algorithms with the minimum possible energy consumption, while assuming that the job start times are already decided to meet the Service Level Agreements (SLAs); and ii) a coordinated cooling-aware job placement and cooling management algorithm, Highest Thermostat Setting (HTS), is developed. HTS is aware of dynamic behavior of the Computer Room Air Conditioner (CRAC) units and places the jobs in a way to reduce the cooling demands from the CRACs. Dynamic updates of the CRAC thermostat settings based on the cooling demands can enable a reduction in energy consumption. Simulation results based on power measurements and job traces from the ASU HPC data center show that: i) HTS reduces the SP-EIR by 15% compared to LRH, a thermal-aware spatial scheduling algorithm; and ii) in conjunction with FCFS-Backfill, HTS increases the throughput per unit energy by 6.89% and 5.56%, respectively, over LRH and MTDP (an energy-effcient spatial scheduling algorithm with server consolidation).
绿色高性能计算数据中心的冷感知和热感知工作负载布局
高性能计算(HPC)数据中心正变得越来越密集;其运行的相关功率密度和能耗正在增加。高达一半的总能量用于冷却数据中心;绿化数据中心运营以减少计算和冷却能源是势在必行的。为此:i)在假设作业开始时间已经确定以满足服务水平协议(sla)的前提下,通过将各算法产生的总(计算+冷却)能耗与可能的最小能耗进行比较,分析空间作业调度(又称作业安置)算法的能源无效率比(SP-EIR);ii)开发了一种协调的冷却感知工作安置和冷却管理算法,即最高恒温器设置(HTS)。HTS了解机房空调(CRAC)机组的动态行为,并安排工作以减少CRAC的冷却需求。基于冷却需求的CRAC恒温器设置的动态更新可以减少能耗。基于ASU HPC数据中心功率测量和作业轨迹的仿真结果表明:与热感知空间调度算法LRH相比,HTS将SP-EIR降低了15%;与fcfs -回填相结合,HTS比LRH和MTDP(一种具有服务器整合的节能空间调度算法)分别提高了6.89%和5.56%的单位能量吞吐量。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信