进程间通信效率对高性能分布式科学计算的影响评价

Ehsan Mousavi Khaneghah, S. L. Mirtaheri, M. Sharifi
{"title":"进程间通信效率对高性能分布式科学计算的影响评价","authors":"Ehsan Mousavi Khaneghah, S. L. Mirtaheri, M. Sharifi","doi":"10.1109/EUC.2008.11","DOIUrl":null,"url":null,"abstract":"Scientific applications like weather forecasting require high performance and fast response time. But this ideal requirement has always been constrained by peculiarities of underlying platforms specially distributed platforms. One such constraint is the efficiency of communication between geographically dispersed and physically distributed processes running these applications, that is the efficiency of inter process communication (IPC) mechanisms. This paper provides hard evidence that an operating system kernel-level implementation of IPC on multi-computers reduces the execution time of a weather forecasting model by nearly half on average compared to when the IPC mechanism is implemented at library level. A well known non-hydrostatic version of the Penn state/NCAR mesoscale model, called MM5, is executed on a networked cluster. The performance of MM5 is measured with two distributed implementations of IPC, a kernel-level implementation called DIPC2006 and a renowned library level implementation called MPI. It is both shown how and argued why the performance of MM5 on a DIPC2006 configured cluster is by far better than its performance on an MPI configured similar cluster. Even ignoring the favorable points of kernel-level implementations, like safety, privilege, reliability, and primitiveness, the insight is twofold. Scientist may look for more efficient distributed implementations of IPC to run their simulations faster, and computer engineers may try harder to develop more efficient distributed implementations of IPC for scientists.","PeriodicalId":430277,"journal":{"name":"2008 IEEE/IFIP International Conference on Embedded and Ubiquitous Computing","volume":"1 2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Evaluating the Effect of Inter Process Communication Efficiency on High Performance Distributed Scientific Computing\",\"authors\":\"Ehsan Mousavi Khaneghah, S. L. Mirtaheri, M. Sharifi\",\"doi\":\"10.1109/EUC.2008.11\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Scientific applications like weather forecasting require high performance and fast response time. But this ideal requirement has always been constrained by peculiarities of underlying platforms specially distributed platforms. One such constraint is the efficiency of communication between geographically dispersed and physically distributed processes running these applications, that is the efficiency of inter process communication (IPC) mechanisms. This paper provides hard evidence that an operating system kernel-level implementation of IPC on multi-computers reduces the execution time of a weather forecasting model by nearly half on average compared to when the IPC mechanism is implemented at library level. A well known non-hydrostatic version of the Penn state/NCAR mesoscale model, called MM5, is executed on a networked cluster. The performance of MM5 is measured with two distributed implementations of IPC, a kernel-level implementation called DIPC2006 and a renowned library level implementation called MPI. It is both shown how and argued why the performance of MM5 on a DIPC2006 configured cluster is by far better than its performance on an MPI configured similar cluster. Even ignoring the favorable points of kernel-level implementations, like safety, privilege, reliability, and primitiveness, the insight is twofold. Scientist may look for more efficient distributed implementations of IPC to run their simulations faster, and computer engineers may try harder to develop more efficient distributed implementations of IPC for scientists.\",\"PeriodicalId\":430277,\"journal\":{\"name\":\"2008 IEEE/IFIP International Conference on Embedded and Ubiquitous Computing\",\"volume\":\"1 2 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2008-12-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2008 IEEE/IFIP International Conference on Embedded and Ubiquitous Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/EUC.2008.11\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 IEEE/IFIP International Conference on Embedded and Ubiquitous Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/EUC.2008.11","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

摘要

像天气预报这样的科学应用需要高性能和快速响应时间。但是这种理想的需求总是受到底层平台(特别是分布式平台)特性的限制。其中一个约束是运行这些应用程序的地理上分散的进程和物理上分布的进程之间的通信效率,即进程间通信(IPC)机制的效率。本文提供了确凿的证据表明,与在库级别实现IPC机制相比,在多台计算机上实现操作系统内核级别的IPC机制可将天气预报模型的执行时间平均减少近一半。宾夕法尼亚州立大学/NCAR中尺度模型的一个著名的非流体静力版本,称为MM5,在一个网络集群上执行。MM5的性能是用两个分布式IPC实现来衡量的,一个是称为DIPC2006的内核级实现,另一个是称为MPI的著名库级实现。本文展示了MM5在配置了DIPC2006的集群上的性能如何远远好于其在配置了MPI的类似集群上的性能,并对此进行了论证。即使忽略内核级实现的优点,如安全性、特权、可靠性和原语性,这种见解也是双重的。科学家可能会寻找更有效的IPC分布式实现来更快地运行他们的模拟,计算机工程师可能会更努力地为科学家开发更有效的IPC分布式实现。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Evaluating the Effect of Inter Process Communication Efficiency on High Performance Distributed Scientific Computing
Scientific applications like weather forecasting require high performance and fast response time. But this ideal requirement has always been constrained by peculiarities of underlying platforms specially distributed platforms. One such constraint is the efficiency of communication between geographically dispersed and physically distributed processes running these applications, that is the efficiency of inter process communication (IPC) mechanisms. This paper provides hard evidence that an operating system kernel-level implementation of IPC on multi-computers reduces the execution time of a weather forecasting model by nearly half on average compared to when the IPC mechanism is implemented at library level. A well known non-hydrostatic version of the Penn state/NCAR mesoscale model, called MM5, is executed on a networked cluster. The performance of MM5 is measured with two distributed implementations of IPC, a kernel-level implementation called DIPC2006 and a renowned library level implementation called MPI. It is both shown how and argued why the performance of MM5 on a DIPC2006 configured cluster is by far better than its performance on an MPI configured similar cluster. Even ignoring the favorable points of kernel-level implementations, like safety, privilege, reliability, and primitiveness, the insight is twofold. Scientist may look for more efficient distributed implementations of IPC to run their simulations faster, and computer engineers may try harder to develop more efficient distributed implementations of IPC for scientists.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信