Java中的大规模并行计算

V. Getov, S. Flynn-Hummel, S. Mintchev, T. Ngo
{"title":"Java中的大规模并行计算","authors":"V. Getov, S. Flynn-Hummel, S. Mintchev, T. Ngo","doi":"10.1109/MPPM.1997.715967","DOIUrl":null,"url":null,"abstract":"Although Java was not specifically designed for the computationally intensive numeric applications that are the typical fodder of highly parallel machines, its widespread popularity and portability make it an interesting candidate vehicle for massively parallel programming. With the advent of high-performance optimizing Java compilers, the open question is: How can Java programs best exploit massive parallelism? The authors have been contemplating this question via libraries of Java-routines for specifying and coordinating parallel codes. It would be most desirable to have these routines written in 100%-Pure Java; however, a more expedient solution is to provide Java wrappers (stubs) to existing parallel coordination libraries, such as MPI. MPI is an attractive alternative, as like Java, it is portable. We discuss both approaches here. In undertaking this study, we have also identified some minor modifications of the current language specification that would make 100%-Pure Java parallel programming more natural.","PeriodicalId":217385,"journal":{"name":"Proceedings. Third Working Conference on Massively Parallel Programming Models (Cat. No.97TB100228)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1997-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Massively parallel computing in Java\",\"authors\":\"V. Getov, S. Flynn-Hummel, S. Mintchev, T. Ngo\",\"doi\":\"10.1109/MPPM.1997.715967\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Although Java was not specifically designed for the computationally intensive numeric applications that are the typical fodder of highly parallel machines, its widespread popularity and portability make it an interesting candidate vehicle for massively parallel programming. With the advent of high-performance optimizing Java compilers, the open question is: How can Java programs best exploit massive parallelism? The authors have been contemplating this question via libraries of Java-routines for specifying and coordinating parallel codes. It would be most desirable to have these routines written in 100%-Pure Java; however, a more expedient solution is to provide Java wrappers (stubs) to existing parallel coordination libraries, such as MPI. MPI is an attractive alternative, as like Java, it is portable. We discuss both approaches here. In undertaking this study, we have also identified some minor modifications of the current language specification that would make 100%-Pure Java parallel programming more natural.\",\"PeriodicalId\":217385,\"journal\":{\"name\":\"Proceedings. Third Working Conference on Massively Parallel Programming Models (Cat. No.97TB100228)\",\"volume\":\"33 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1997-11-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings. Third Working Conference on Massively Parallel Programming Models (Cat. No.97TB100228)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/MPPM.1997.715967\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings. Third Working Conference on Massively Parallel Programming Models (Cat. No.97TB100228)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MPPM.1997.715967","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

摘要

尽管Java并不是专门为计算密集型数值应用程序而设计的,但它的广泛流行和可移植性使其成为大规模并行编程的有趣候选工具。随着高性能优化Java编译器的出现,一个开放的问题是:Java程序如何才能最好地利用大规模并行性?作者一直在通过指定和协调并行代码的java例程库来考虑这个问题。用100%纯Java编写这些例程是最理想的;然而,更方便的解决方案是为现有的并行协调库(如MPI)提供Java包装器(存根)。MPI是一个有吸引力的替代方案,就像Java一样,它是可移植的。我们在这里讨论这两种方法。在进行这项研究时,我们还确定了对当前语言规范的一些小修改,这些修改将使100%纯Java并行编程更加自然。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Massively parallel computing in Java
Although Java was not specifically designed for the computationally intensive numeric applications that are the typical fodder of highly parallel machines, its widespread popularity and portability make it an interesting candidate vehicle for massively parallel programming. With the advent of high-performance optimizing Java compilers, the open question is: How can Java programs best exploit massive parallelism? The authors have been contemplating this question via libraries of Java-routines for specifying and coordinating parallel codes. It would be most desirable to have these routines written in 100%-Pure Java; however, a more expedient solution is to provide Java wrappers (stubs) to existing parallel coordination libraries, such as MPI. MPI is an attractive alternative, as like Java, it is portable. We discuss both approaches here. In undertaking this study, we have also identified some minor modifications of the current language specification that would make 100%-Pure Java parallel programming more natural.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信