{"title":"The virtual-time data-parallel machine","authors":"S. Shen, L. Kleinrock","doi":"10.1109/FMPC.1992.234906","DOIUrl":null,"url":null,"abstract":"The authors propose the virtual-time data-parallel machine to execute SIMD (single instruction multiple data) programs asynchronously. They first illustrate how asynchronous execution is more efficient than synchronous execution. For a simple model, they show that asynchronous execution outperforms synchronous execution roughly by a factor of (ln N), where N is the number of processors in the system. They then explore how to execute SIMD programs asynchronously without violating the SIMD semantics. They design a first in, first out (FIFO) priority cache, one for each processing element, to record the recent history of all variables. The cache, which is stacked between the processor and the memory, supports asynchronous execution in hardware efficiently and preserves the SIMD semantics of the software transparently. Analysis and simulation results indicate that the virtual-time data-parallel machine can achieve linear speed-up for computation-intensive data-parallel programs when the number of processors is large.<<ETX>>","PeriodicalId":117789,"journal":{"name":"[Proceedings 1992] The Fourth Symposium on the Frontiers of Massively Parallel Computation","volume":"21 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"[Proceedings 1992] The Fourth Symposium on the Frontiers of Massively Parallel Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/FMPC.1992.234906","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7
Abstract
The authors propose the virtual-time data-parallel machine to execute SIMD (single instruction multiple data) programs asynchronously. They first illustrate how asynchronous execution is more efficient than synchronous execution. For a simple model, they show that asynchronous execution outperforms synchronous execution roughly by a factor of (ln N), where N is the number of processors in the system. They then explore how to execute SIMD programs asynchronously without violating the SIMD semantics. They design a first in, first out (FIFO) priority cache, one for each processing element, to record the recent history of all variables. The cache, which is stacked between the processor and the memory, supports asynchronous execution in hardware efficiently and preserves the SIMD semantics of the software transparently. Analysis and simulation results indicate that the virtual-time data-parallel machine can achieve linear speed-up for computation-intensive data-parallel programs when the number of processors is large.<>