HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.最新文献

筛选
英文 中文
Automatic dynamic run-time optical network reservations 自动动态运行时光网络预约
J. Lange, Ananth I. Sundararaj, P. Dinda
{"title":"Automatic dynamic run-time optical network reservations","authors":"J. Lange, Ananth I. Sundararaj, P. Dinda","doi":"10.1109/HPDC.2005.1520969","DOIUrl":"https://doi.org/10.1109/HPDC.2005.1520969","url":null,"abstract":"Optical networking may dramatically change high performance distributed computing. One reason is that optical networks can support provisioning dynamically configurable lightpaths, a form of circuit switching, through reservations. However, to use it (and all other network reservation mechanisms), the user or developer must modify the application. We present a system, VRESERVE, that automatically and dynamically creates network reservation requests based on the inferred network demands of running distributed and/or parallel applications with no modification to the application or operating system, and no input from the user or developer. Our execution model is a collection of virtual machines interconnected by an overlay network. The overlay network infers application demands, providing a dynamic run-time assessment of the application's topology and traffic load matrix. We then reserve lightpaths corresponding to the topology and use the overlay to forward virtual network traffic over them. We evaluate our system on the OMNInet network.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"26 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126136921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Building self-adapting services using service-specific knowledge 使用特定于服务的知识构建自适应服务
An-Cheng Huang, P. Steenkiste
{"title":"Building self-adapting services using service-specific knowledge","authors":"An-Cheng Huang, P. Steenkiste","doi":"10.1109/HPDC.2005.1520931","DOIUrl":"https://doi.org/10.1109/HPDC.2005.1520931","url":null,"abstract":"With the advances in middleware and Web services technologies, network sendees are evolving from simple client-sender applications to self-configuring services that can compose primitive components distributed in the Internet into a value-added service configuration that provides rich functionalities to users. A resulting research problem is how to continuously adapt such composite service configurations at run time in order to cope with the increasingly dynamic and heterogeneous network environments and computing platforms. In this paper, we propose a self-adaptation architecture that allows service developers to specify their service-specific adaptation knowledge as \"externalized\" adaptation strategies. These adaptation strategies are used by a general, shared adaptation framework to perform run-time adaptation operations that automatically incorporate service-specific knowledge. In addition to the strategies, we also identify another aspect of adaptation knowledge that is not addressed by previous solutions: adaptation coordination. Our framework provides integrated support for the specification and execution of both aspects of developers' adaptation knowledge.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122127893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Generosity and gluttony in GEMS: grid enabled molecular simulations GEMS中的慷慨与暴饮暴食:网格分子模拟
J. Wozniak, P. Brenner, D. Thain, A. Striegel, J. Izaguirre
{"title":"Generosity and gluttony in GEMS: grid enabled molecular simulations","authors":"J. Wozniak, P. Brenner, D. Thain, A. Striegel, J. Izaguirre","doi":"10.1109/HPDC.2005.1520959","DOIUrl":"https://doi.org/10.1109/HPDC.2005.1520959","url":null,"abstract":"Biomolecular simulations produce more output data than can be managed effectively by traditional computing systems. Researchers need distributed systems that allow the pooling of resources, the sharing of simulation data, and the reliable publication of both tentative and final results. To address this need, we have designed GEMS, a system that enables biomolecular researchers to store, search, and share large scale simulation data. The primary design problem is striking a balance between generosity and gluttony. On one hand, storage providers wish to be generous and share resources with their collaborators. On the other hand, an unchecked data producer can be gluttonous and easily replicate data unnecessarily until it fills all available space. To balance generosity and gluttony, GEMS allows both storage providers and data producers to state and enforce policies on the consumption of storage and the replication of data. By taking advantage of known properties of simulation data, the system is able to distinguish between high value final results that must be preserved and low value intermediate results that can be deleted and regenerated if necessary. We have built a prototype of GEMS on a cluster of workstations and demonstrate its ability to store new data, to replicate within policy limits, and to recover from failures.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129187730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Cluster delegation: high-performance, fault-tolerant data sharing in NFS 集群委托:在NFS中实现高性能、容错的数据共享
Alexandros Batsakis, R. Burns
{"title":"Cluster delegation: high-performance, fault-tolerant data sharing in NFS","authors":"Alexandros Batsakis, R. Burns","doi":"10.1109/HPDC.2005.1520942","DOIUrl":"https://doi.org/10.1109/HPDC.2005.1520942","url":null,"abstract":"We present cluster delegation, an enhancement to the NFSv4 file system, that improves both performance and recoverability in computing clusters. Cluster delegation allows data sharing among clients by extending the NFSv4 delegation model so that multiple clients manage a single file without interacting with the server. Based on cluster delegation, we implement a fast commit primitive, cooperative caching, and the ability to recover the uncommitted updates of a failed computer. Cluster delegation supports both read and write operations in the cooperative cache, while preserving the consistency guarantees of NFSv4. We have implemented cluster delegation by modifying the Linux NFSv4 client and show that it improves client performance and reduces server load by more than half.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126451765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Performance analysis of a file catalog for the LHC computing grid LHC计算网格的文件编目性能分析
J. Baud, J. Casey, S. Lemaitre, C. Nicholson
{"title":"Performance analysis of a file catalog for the LHC computing grid","authors":"J. Baud, J. Casey, S. Lemaitre, C. Nicholson","doi":"10.1109/HPDC.2005.1520941","DOIUrl":"https://doi.org/10.1109/HPDC.2005.1520941","url":null,"abstract":"The Large Hadron Collider (LHC) at CERN, the European Organization for Nuclear Research, needs to produce unprecedented volumes of data when it starts operation in 2007. To provide for its computational needs, the LHC computing grid (LCG) should be deployed as a worldwide computational grid service, providing the middleware upon which the physics analysis for the LHC can be carried out. In 2003, versions of this middleware were deployed which were based on the middleware produced by the European Data Grid project (EDG). In 2004 the LCG-2 release, which consisted of the EDG middleware with some minor modifications, was deployed for use by the LHC experiments. A series of data challenges by these experiments were the first real experiment production use of LCG. During the course of the data challenges, many issues and problems were exposed which had not shown up in more limited tests. The deployment, service and development teams worked closely with the experiments to understand these issues and while some of the problems were solved during the data challenges, others exposed fundamental problems with the middleware as deployed in LCG-2. One of these fundamental problems was the performance under real load of the catalog component provided by EDG, the replica location service. To solve these problems a new component was designed, the LCG file catalog (LFC). The LFC moves away from the replica location service model used in previous LCG releases, towards a hierarchical file system model which is more like a UNIX file system. It also adds missing functionality which was requested by the experiments. This paper presents the architecture and implementation of the LFC and evaluates it in a series of performance tests, with up to forty million entries and one hundred requesting threads from multiple clients. The results show good scalability up to the limits of these tests, and compare favourably with other grid catalog implementations.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121961570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Design and implementation tradeoffs for wide-area resource discovery 广域资源发现的设计和实现权衡
Jeannie R. Albrecht, David L. Oppenheimer, Amin Vahdat, D. Patterson
{"title":"Design and implementation tradeoffs for wide-area resource discovery","authors":"Jeannie R. Albrecht, David L. Oppenheimer, Amin Vahdat, D. Patterson","doi":"10.1145/1391949.1391952","DOIUrl":"https://doi.org/10.1145/1391949.1391952","url":null,"abstract":"This paper describes the design and implementation of SWORD, a scalable resource discovery service for wide-area distributed systems. In contrast to previous systems, SWORD allows users to describe desired resources as a topology of interconnected groups with required intragroup, intergroup, and per-node characteristics, along with the utility that the application derives from various ranges of values of those characteristics. This design gives users the flexibility to find geographically distributed resources for applications that are sensitive to both node and network characteristics, and allows the system to rank acceptable configurations based on their quality for that application. We explore a variety of architectures to deliver SWORD's functionality in a scalable and highly-available manner. A 1000-node ModelNet evaluation using a workload of measurements collected from PlanetLab shows that an architecture based on 4-node server cluster sites at network peering facilities outperforms a decentralized DHT-based resource discovery infrastructure for all but the smallest number of sites. While such a centralized architecture shows significant promise, we find that our decentralized implementation, both in emulation and running continuously on over 200 PlanetLab nodes, performs well while benefiting from the DHT's self-healing properties.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"1049 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131823319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 227
Techniques for tuning workflows in cluster environments 在集群环境中调优工作流的技术
B. Tierney, D. Gunter
{"title":"Techniques for tuning workflows in cluster environments","authors":"B. Tierney, D. Gunter","doi":"10.1109/HPDC.2005.1520986","DOIUrl":"https://doi.org/10.1109/HPDC.2005.1520986","url":null,"abstract":"An important class of parallel processing jobs on clusters today are workflow-based applications that process large amounts of data in parallel. Traditional cluster performance tools are designed for tightly coupled parallel jobs, and not as effective for this type of application. We describe how the NetLogger Toolkit methodology is more appropriate for this class of cluster computing, and describe our new automatic workflow anomaly detection component. We also describe how this methodology is being used by the Nearby Supernova Factory (SNfactory) project at Lawrence Berkeley National Laboratory.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133407840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A model for cooperative federation of distributed clusters 分布式集群的合作联合模型
R. Ranjan, R. Buyya, A. Harwood
{"title":"A model for cooperative federation of distributed clusters","authors":"R. Ranjan, R. Buyya, A. Harwood","doi":"10.1109/HPDC.2005.1520982","DOIUrl":"https://doi.org/10.1109/HPDC.2005.1520982","url":null,"abstract":"Interest in grid computing has grown significantly over the past five years. Management of distributed cluster resources is a key issue in grid computing. Central to management of resources is the effectiveness of resource allocation, as it determines the overall utility of the system. In this paper, we propose a new grid system that consists of grid federation agents which couple together distributed cluster resources to enable a cooperative environment.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131328562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Genetic algorithm based automatic data partitioning scheme for HPF 基于遗传算法的HPF自动数据分区方案
S. Anand, Y. Srikant
{"title":"Genetic algorithm based automatic data partitioning scheme for HPF","authors":"S. Anand, Y. Srikant","doi":"10.1109/HPDC.2005.1520979","DOIUrl":"https://doi.org/10.1109/HPDC.2005.1520979","url":null,"abstract":"The performance of a parallel program depends largely on its data partitions. So a good data partitioning scheme is the need of the time. However it is very difficult to arrive at a good solution as the number of possible data partitions for a given real life program is exponential in the size of the program. We present a heuristic technique for automatic data partitioning for HPF. Our approach is based on genetic algorithms and is very simple, yet very efficient to quickly find appropriate data partitions even for large programs with large number of alternatives for data distribution. It makes use of both static as well as dynamic data distribution with the main aim of reducing the overall execution time of the entire program.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114640254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HMPI - hybrid MPI HMPI混合MPI
F. I. Massetto, Augusto Mendes Gomes Junior, L. Sato
{"title":"HMPI - hybrid MPI","authors":"F. I. Massetto, Augusto Mendes Gomes Junior, L. Sato","doi":"10.1109/HPDC.2005.1520987","DOIUrl":"https://doi.org/10.1109/HPDC.2005.1520987","url":null,"abstract":"This paper presents the HMPI, a runtime system to integrate several MPI implementations, used to develop high performance applications that must run both in nodes with several operating systems and clusters of clusters infrastructures. HMPI has two different approaches to achieve this integration, using connection pools and HMPI daemons.","PeriodicalId":120564,"journal":{"name":"HPDC-14. Proceedings. 14th IEEE International Symposium on High Performance Distributed Computing, 2005.","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115637596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信