Workshop on Exascale MPI最新文献

筛选
英文 中文
Adaptive transport service selection for MPI with InfiniBand network 基于ib网络的MPI自适应传输服务选择
Workshop on Exascale MPI Pub Date : 2015-11-15 DOI: 10.1145/2831129.2831132
Masamichi Takagi, Norio Yamaguchi, Balazs Gerofi, A. Hori, Y. Ishikawa
{"title":"Adaptive transport service selection for MPI with InfiniBand network","authors":"Masamichi Takagi, Norio Yamaguchi, Balazs Gerofi, A. Hori, Y. Ishikawa","doi":"10.1145/2831129.2831132","DOIUrl":"https://doi.org/10.1145/2831129.2831132","url":null,"abstract":"We propose a method which adaptively selects InfiniBand transport services used for source and destination peers to improve performance while limiting memory consumption of the MPI library. There are two major choices of IB transport services available, i.e., Reliable Connection (RC) and Dynamically Connected (DC), each of which is selected for each pair of source peer and destination peer. RC is faster than DC for all communication patterns except for the case where there are many active RCs. It also consumes a lot of memory when there are many processes. DC, on the other hand, consumes less memory than RC but its performance drops when sending messages to different destinations or when many DCs sends a message to the same destination DC. Therefore, the library should find the best mapping of RCs and DCs to pairs of source peer and destination peer according to the communication pattern of the application. Our method finds a good mapping by comparing potential latency benefits for candidate mappings. It achieves 13%--19% latency reduction when compared to the methods using only DCs in micro-benchmarks representing communication patterns problematic to RC or DC with 64 processes.","PeriodicalId":417011,"journal":{"name":"Workshop on Exascale MPI","volume":"30 15","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132914584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Practical resilient cases for FA-MPI, a transactional fault-tolerant MPI FA-MPI的实际弹性案例,一个事务性容错MPI
Workshop on Exascale MPI Pub Date : 2015-11-15 DOI: 10.1145/2831129.2831130
Amin Hassani, A. Skjellum, P. Bangalore, R. Brightwell
{"title":"Practical resilient cases for FA-MPI, a transactional fault-tolerant MPI","authors":"Amin Hassani, A. Skjellum, P. Bangalore, R. Brightwell","doi":"10.1145/2831129.2831130","DOIUrl":"https://doi.org/10.1145/2831129.2831130","url":null,"abstract":"MPI is insufficient when confronting failures. FA-MPI (Fault-Aware MPI) provides extensions to the MPI standard designed to enable data-parallel applications to achieve resilience without sacrificing scalability. FA-MPI introduces transactions as a novel extension to the MPI message-passing model. Transactions support failure detection, isolation, mitigation, and recovery via application-driven policies. To achieve maximum achievable performance of modern machines, overlapping communication and I/O with computation through non-blocking operations is of growing importance. Therefore, we emphasize fault-tolerant, non-blocking communication operations plus a set of nestable lightweight transactional TryBlock API extensions able to exploit system and application hierarchy. This strategy enables applications to run to completion with higher probability than nominally. We modified two proxy applications---MiniFE and LULESH---by adding FA-MPI semantics to them. Finally we present performance and overhead results for 1K MPI processes.","PeriodicalId":417011,"journal":{"name":"Workshop on Exascale MPI","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130981838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
A data streaming model in MPI MPI中的数据流模型
Workshop on Exascale MPI Pub Date : 2015-11-15 DOI: 10.1145/2831129.2831131
I. Peng, S. Markidis, E. Laure, Daniel J. Holmes, Mark Bull
{"title":"A data streaming model in MPI","authors":"I. Peng, S. Markidis, E. Laure, Daniel J. Holmes, Mark Bull","doi":"10.1145/2831129.2831131","DOIUrl":"https://doi.org/10.1145/2831129.2831131","url":null,"abstract":"Data streaming model is an effective way to tackle the challenge of data-intensive applications. As traditional HPC applications generate large volume of data and more data-intensive applications move to HPC infrastructures, it is necessary to investigate the feasibility of combining message-passing and streaming programming models. MPI, the de facto standard for programming on HPC, cannot intuitively express the communication pattern and the functional operations required in streaming models. In this work, we designed and implemented a data streaming library MPIStream atop MPI to allocate data producers and consumers, to stream data continuously or irregularly and to process data at run-time. In the same spirit as the STREAM benchmark, we developed a parallel stream benchmark to measure data processing rate. The performance of the library largely depends on the size of the stream element, the number of data producers and consumers and the computational intensity of processing one stream element. With 2,048 data producers and 2,048 data consumers in the parallel benchmark, MPIStream achieved 200 GB/s processing rate on a Blue Gene/Q supercomputer. We illustrate that a streaming library for HPC applications can effectively enable irregular parallel I/O, application monitoring and threshold collective operations.","PeriodicalId":417011,"journal":{"name":"Workshop on Exascale MPI","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115300397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Overtime: a tool for analyzing performance variation due to network interference 加班:分析网络干扰导致的性能变化的工具
Workshop on Exascale MPI Pub Date : 2015-11-15 DOI: 10.1145/2831129.2831133
Ryan E. Grant, K. Pedretti, A. Gentile
{"title":"Overtime: a tool for analyzing performance variation due to network interference","authors":"Ryan E. Grant, K. Pedretti, A. Gentile","doi":"10.1145/2831129.2831133","DOIUrl":"https://doi.org/10.1145/2831129.2831133","url":null,"abstract":"Shared networks create unique challenges in obtaining consistent performance across jobs for large systems when not using exclusive system-wide allocations. In order to provide good system utilization, resource managers allocate system space to multiple jobs. These multiple independent node allocations can interfere with each other through their shared network. This work provides a method of observing and measuring the impact of network contention due to interference from other jobs through a continually running benchmark application and the use of network performance counters. This is the first work to measure network interference using specially designed benchmarks and network performance counters.","PeriodicalId":417011,"journal":{"name":"Workshop on Exascale MPI","volume":"31 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125758962","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Preparing for exascale: modeling MPI for many-core systems using fine-grain queues 为百亿亿级做准备:使用细粒度队列为多核系统建模MPI
Workshop on Exascale MPI Pub Date : 2015-11-15 DOI: 10.1145/2831129.2831134
P. Bridges, Matthew G. F. Dosanjh, Ryan E. Grant, A. Skjellum, Shane Farmer, R. Brightwell
{"title":"Preparing for exascale: modeling MPI for many-core systems using fine-grain queues","authors":"P. Bridges, Matthew G. F. Dosanjh, Ryan E. Grant, A. Skjellum, Shane Farmer, R. Brightwell","doi":"10.1145/2831129.2831134","DOIUrl":"https://doi.org/10.1145/2831129.2831134","url":null,"abstract":"This paper presents a fine-grain queueing model of MPI point-to-point messaging performance for use in the design and analysis of current and future large-scale computing systems. In particular, the model seeks to capture key performance behavior of MPI communication on many-core systems. We demonstrate that this model encompasses key MPI performance characteristics, such as short/long protocol and offload/onload protocol tradeoffs, and demonstrate its use in predicting the potential impact of architectural and software changes for many-core systems on communication performance. In addition, we also discuss the limitations of this model and potential directions for enhancing its fidelity.","PeriodicalId":417011,"journal":{"name":"Workshop on Exascale MPI","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130952822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信