Mohsin Khan, N. Priyanka, W. Ahmed, N. Radhika, M. Pavithra, K. Parimala
{"title":"Understanding source-to-source transformations for frequent porting of applications on changing cloud architectures","authors":"Mohsin Khan, N. Priyanka, W. Ahmed, N. Radhika, M. Pavithra, K. Parimala","doi":"10.1109/PDGC.2014.7030769","DOIUrl":null,"url":null,"abstract":"Writing code for heterogeneous architectures with processors and accelerators from multiple vendors from scratch or translating existing serial code, a lot of effort and investment will be required from the application developer. This problem will become more prominent when HPC applications are moved into the Cloud as Cloud providers frequently update their architectures to keep with market trends. In these scenarios, automatic parallelization tools will definitely have an important role to play. An important constituent of these tools would be the ability to perform pertinent domain decomposition of the serial code to maximize utilization of the available computational elements. One of the first steps in this direction is to understand the role of the number and type of computational element in a heterogeneous architecture to the overall performance of an application. This paper presents observations made on architectures with different types and number of computational elements using two case studies on five different architectures with different types and number of computational elements. Results show that the perceived speedup and actual speedup are not related.","PeriodicalId":311953,"journal":{"name":"2014 International Conference on Parallel, Distributed and Grid Computing","volume":"240 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2014 International Conference on Parallel, Distributed and Grid Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/PDGC.2014.7030769","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Writing code for heterogeneous architectures with processors and accelerators from multiple vendors from scratch or translating existing serial code, a lot of effort and investment will be required from the application developer. This problem will become more prominent when HPC applications are moved into the Cloud as Cloud providers frequently update their architectures to keep with market trends. In these scenarios, automatic parallelization tools will definitely have an important role to play. An important constituent of these tools would be the ability to perform pertinent domain decomposition of the serial code to maximize utilization of the available computational elements. One of the first steps in this direction is to understand the role of the number and type of computational element in a heterogeneous architecture to the overall performance of an application. This paper presents observations made on architectures with different types and number of computational elements using two case studies on five different architectures with different types and number of computational elements. Results show that the perceived speedup and actual speedup are not related.