Enabling parallel scientific applications with workflow tools

Adam Lathers, Mei-Hui Su, A. Kulungowski, A. Lin, Gaurang Mehta, S. Peltier, E. Deelman, Mark Ellisman
{"title":"Enabling parallel scientific applications with workflow tools","authors":"Adam Lathers, Mei-Hui Su, A. Kulungowski, A. Lin, Gaurang Mehta, S. Peltier, E. Deelman, Mark Ellisman","doi":"10.1109/CLADE.2006.1652055","DOIUrl":null,"url":null,"abstract":"Electron tomography is a powerful tool for deriving three-dimensional (3D) structural information about biological systems within the spatial scale spanning 1 nm3 and 10 mm3. With this technique, it is possible to derive detailed models of sub-cellular components such as organelles and synaptic complexes and to resolve the 3D distribution of their protein constituents in situ. Due in part to exponentially growing raw data-sizes, there continues to be a need for the increased integration of high-performance computing (HPC) and grid technologies with traditional electron tomography processes to provide faster data processing throughput. This is increasingly relevant because emerging mathematical algorithms that provide better data fidelity are more computationally intensive for larger raw data sizes. Progress has been made towards the transparent use of HPC and grid tools for launching scientific applications without passing on the necessary administrative overhead and complexity (resource administration, authentication, scheduling, data delivery) to the non-computer scientist end-user. There is still a need, however, to simplify the use of these tools for applications developers who are developing novel algorithms for computation. Here we describe the architecture of the Telescience project (http://telescience.ucsd.edu), specifically the use of layered workflow technologies to parallelize and execute scientific codes across a distributed and heterogeneous computational resource pool (including resources from the TeraGrid and OptlPuter projects) without the need for the application developer to understand the intricacies of the grid","PeriodicalId":299480,"journal":{"name":"2006 IEEE Challenges of Large Applications in Distributed Environments","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2006-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"25","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2006 IEEE Challenges of Large Applications in Distributed Environments","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CLADE.2006.1652055","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 25

Abstract

Electron tomography is a powerful tool for deriving three-dimensional (3D) structural information about biological systems within the spatial scale spanning 1 nm3 and 10 mm3. With this technique, it is possible to derive detailed models of sub-cellular components such as organelles and synaptic complexes and to resolve the 3D distribution of their protein constituents in situ. Due in part to exponentially growing raw data-sizes, there continues to be a need for the increased integration of high-performance computing (HPC) and grid technologies with traditional electron tomography processes to provide faster data processing throughput. This is increasingly relevant because emerging mathematical algorithms that provide better data fidelity are more computationally intensive for larger raw data sizes. Progress has been made towards the transparent use of HPC and grid tools for launching scientific applications without passing on the necessary administrative overhead and complexity (resource administration, authentication, scheduling, data delivery) to the non-computer scientist end-user. There is still a need, however, to simplify the use of these tools for applications developers who are developing novel algorithms for computation. Here we describe the architecture of the Telescience project (http://telescience.ucsd.edu), specifically the use of layered workflow technologies to parallelize and execute scientific codes across a distributed and heterogeneous computational resource pool (including resources from the TeraGrid and OptlPuter projects) without the need for the application developer to understand the intricacies of the grid
使用工作流工具启用并行科学应用程序
电子断层扫描是一种强大的工具,可以在1nm3和10mm3的空间尺度内获得关于生物系统的三维(3D)结构信息。利用这种技术,可以推导出亚细胞成分(如细胞器和突触复合物)的详细模型,并在原位解决其蛋白质成分的3D分布。部分由于原始数据大小呈指数级增长,因此仍然需要将高性能计算(HPC)和网格技术与传统电子断层扫描过程集成在一起,以提供更快的数据处理吞吐量。这是越来越重要的,因为提供更好的数据保真度的新兴数学算法对于更大的原始数据规模来说计算更加密集。在透明地使用HPC和网格工具来启动科学应用程序方面已经取得了进展,而无需将必要的管理开销和复杂性(资源管理、认证、调度、数据交付)传递给非计算机科学家的最终用户。然而,对于正在开发新的计算算法的应用程序开发人员来说,仍然需要简化这些工具的使用。在这里,我们描述了Telescience项目的架构(http://telescience.ucsd.edu),特别是使用分层工作流技术在分布式和异构计算资源池(包括来自TeraGrid和OptlPuter项目的资源)上并行化和执行科学代码,而不需要应用程序开发人员理解网格的复杂性
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信