{"title":"mpi集群中多核、多处理器节点的编程影响","authors":"H. Pourreza, P. Graham","doi":"10.1109/HPCS.2007.23","DOIUrl":null,"url":null,"abstract":"Increasingly, multi-core processors, multi-processor nodes and multi-core, multi-processor nodes are finding their way into computer clusters. Clusters built using such nodes are already quite common and, inevitably, will become more so over time. As with any new technology, however, the potential benefits are seldom as easy to attain as we expect them to be. In this paper, we explore three fundamental issues related to the use of multi-core, multi-processor nodes in compute clusters using MPI: inter-communication (messaging) efficiency, cache effects (in particular processor affinity) and initial process distribution. Based on some initial experiments using a subset of the NAS parallel benchmarks running on a small scale cluster with dual core, dual processor nodes, we report results on the impact of these issues. From these results we attempt to extrapolate some simple, guidelines that are likely to be generally applicable for optimizing MPI code running on clusters with multi- core, multi-processor nodes.","PeriodicalId":354520,"journal":{"name":"21st International Symposium on High Performance Computing Systems and Applications (HPCS'07)","volume":"2013 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"9","resultStr":"{\"title\":\"On the Programming Impact ofMulti-Core,Multi-Processor Nodes inMPI Clusters\",\"authors\":\"H. Pourreza, P. Graham\",\"doi\":\"10.1109/HPCS.2007.23\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Increasingly, multi-core processors, multi-processor nodes and multi-core, multi-processor nodes are finding their way into computer clusters. Clusters built using such nodes are already quite common and, inevitably, will become more so over time. As with any new technology, however, the potential benefits are seldom as easy to attain as we expect them to be. In this paper, we explore three fundamental issues related to the use of multi-core, multi-processor nodes in compute clusters using MPI: inter-communication (messaging) efficiency, cache effects (in particular processor affinity) and initial process distribution. Based on some initial experiments using a subset of the NAS parallel benchmarks running on a small scale cluster with dual core, dual processor nodes, we report results on the impact of these issues. From these results we attempt to extrapolate some simple, guidelines that are likely to be generally applicable for optimizing MPI code running on clusters with multi- core, multi-processor nodes.\",\"PeriodicalId\":354520,\"journal\":{\"name\":\"21st International Symposium on High Performance Computing Systems and Applications (HPCS'07)\",\"volume\":\"2013 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2007-05-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"9\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"21st International Symposium on High Performance Computing Systems and Applications (HPCS'07)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/HPCS.2007.23\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"21st International Symposium on High Performance Computing Systems and Applications (HPCS'07)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/HPCS.2007.23","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
On the Programming Impact ofMulti-Core,Multi-Processor Nodes inMPI Clusters
Increasingly, multi-core processors, multi-processor nodes and multi-core, multi-processor nodes are finding their way into computer clusters. Clusters built using such nodes are already quite common and, inevitably, will become more so over time. As with any new technology, however, the potential benefits are seldom as easy to attain as we expect them to be. In this paper, we explore three fundamental issues related to the use of multi-core, multi-processor nodes in compute clusters using MPI: inter-communication (messaging) efficiency, cache effects (in particular processor affinity) and initial process distribution. Based on some initial experiments using a subset of the NAS parallel benchmarks running on a small scale cluster with dual core, dual processor nodes, we report results on the impact of these issues. From these results we attempt to extrapolate some simple, guidelines that are likely to be generally applicable for optimizing MPI code running on clusters with multi- core, multi-processor nodes.