{"title":"A Resource Synchronization Protocol for Multiprocessor Real-Time Systems","authors":"Chia-Mei Chen, S. Tripathi, A. Blackmore","doi":"10.1109/ICPP.1994.44","DOIUrl":"https://doi.org/10.1109/ICPP.1994.44","url":null,"abstract":"We study resource synchronization in multiprocessor hard real-time systems. Specifically, we propose a multiprocessor resource control protocol which allows a job to simultaneously lock multiple global resources, removing a restriction from previous protocols. Allowing nested critical sections may permit a finer granularity of synchronization, increasing parallelism and throughput. All the protocols discussed belong to the class of priority inheritance protocols and rely in some fashion on priority ceilings for global semaphores. The extended protocol prevents deadlock and transitive blocking. We derive bounds for worse case blocking time, and describe sufficient conditions to guarantee that m sets of periodic tasks can be scheduled on an m multiprocessor system.","PeriodicalId":162043,"journal":{"name":"1994 International Conference on Parallel Processing Vol. 3","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132664300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Constant Time Graph and Poset Algorithms on the Reconfigurable Multiple Bus Machine","authors":"J. Trahan, R. Vaidyanathan, C. Subbaraman","doi":"10.1109/ICPP.1994.78","DOIUrl":"https://doi.org/10.1109/ICPP.1994.78","url":null,"abstract":"The Reconfigurable Multiple Bus Machine (RMBM) is a model of parallel computation based on reconfigurable buses. In this paper, vie present constant time RMBM algorithms for a collection of basic, graph problems that include, lowest common ancestors and Euler tour related problems (for trees) and shortest path and connectivity related problems (for general graphs). We also present results for some poset and lattice problems. All algorithms are at least as efficient or more efficient in terms of processors than corresponding PARBUS algorithms.","PeriodicalId":162043,"journal":{"name":"1994 International Conference on Parallel Processing Vol. 3","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132756700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Speculative Computation: Overcoming Communication Delays","authors":"V. Govindan, M. Franklin","doi":"10.1109/ICPP.1994.183","DOIUrl":"https://doi.org/10.1109/ICPP.1994.183","url":null,"abstract":"Communication latencies and delays are a major source of performance degradation in parallel computing systems. It is important to \"mask\" these communication delays by overlapping them with useful computation in order to obtain good parallel performance. This article proposes speculative computation as a technique to mask communication latencies in synchronous iterative algorithms. Processors speculate the contents of messages that are not yet received and perform computation based on the speculated values. When the messages are received, they are compared with the speculated values and, if the error is unacceptable, the resulting computation is corrected or recomputed. If the error is small, the speculated value is accepted and the processor has masked the communication delay. The technique, applied to N-body simulations yielded a performance improvement of up to 34%.","PeriodicalId":162043,"journal":{"name":"1994 International Conference on Parallel Processing Vol. 3","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114864323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"MPI-F: An Efficient Implementation of MPI on IBM-SP1","authors":"H. Franke, P. Hochschild, P. Pattnaik, M. Snir","doi":"10.1109/ICPP.1994.131","DOIUrl":"https://doi.org/10.1109/ICPP.1994.131","url":null,"abstract":"This article introduces MPI-F an efficient implementation of MPI on the IBM-SP1 distributed memory cluster. After discussing the novel and key concepts of MPI and how they relate to an implementation, the MPI-F system architecture is outlined in detail. Although many incorrectly assume that MPI will not be efficient due to its increased functionality, MPI-F performance demonstrates efficiency as good as the best message passing library currently available on the SP1.","PeriodicalId":162043,"journal":{"name":"1994 International Conference on Parallel Processing Vol. 3","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128986738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Efficient Parallel Computation of Projections for Animated Volume Visualization","authors":"Tuang-Kuang Wu, M. Brady","doi":"10.1109/ICPP.1994.102","DOIUrl":"https://doi.org/10.1109/ICPP.1994.102","url":null,"abstract":"In this paper, we parallelize a new efficient approximate projection algorithm for volume rendering. The algorithm is unique in that it computes N different views simultaneously, much faster than these projections could be computed independently. It is thus ideal for computing animated sequences of views. We implement and evaluate the algorithm on a MasPar MP-1. We render a 128x128x128 data set at up to 21 frames per second average on a 1K PE MasPar MP-1 using gray scale volumetric compositing (simpler transformations are computed even faster).","PeriodicalId":162043,"journal":{"name":"1994 International Conference on Parallel Processing Vol. 3","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125532561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Scheduling Strategies for Sparse Cholesky Factorization on a Shared Virtual Memory Parallel Computer","authors":"M. Hahad, J. Erhel, T. Priol","doi":"10.1109/ICPP.1994.177","DOIUrl":"https://doi.org/10.1109/ICPP.1994.177","url":null,"abstract":"To solve a given problem on a distributed memory parallel computer (DMPC), the message passing programming model involves distributing both the data and the computations among the processors. While this can be easily feasible for well structured problems, it can become fairly hard for unstructured ones, like sparse matrix computations, unless you use some runtime support. In this paper, we consider a relatively new approach to implementing the Cholesky factorization on a DMPC, by using a shared virtual memory (SVM). The abstraction of a shared memory on top of a distributed memory allows us to introduce a large-grain factorization algorithm, synchronized with events. Experiments conducted so far show that some scheduling techniques enhance not only the parallelism but the SVM behavior as well, allowing interesting results.","PeriodicalId":162043,"journal":{"name":"1994 International Conference on Parallel Processing Vol. 3","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125974299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Evaluation of a Local Adaptive Protocol for Distributed Discrete Event Simulation","authors":"Donald O. Hamnes, A. Tripathi","doi":"10.1109/ICPP.1994.106","DOIUrl":"https://doi.org/10.1109/ICPP.1994.106","url":null,"abstract":"The performance of a local adaptive protocol for distributed discrete event simulation is evaluated. This protocol allows each process to adapt at runtime to its environment on a per channel basis with the possible range of behavior varying from conservative to optimistic. This evaluation includes a comparison of the performance of the algorithm with that of both a conservative and an optimistic protocol, as well as, an examination of a number of unique characteristics of the algorithm. Results from simulations using a network of workstations on a closed queueing system are presented. The results demonstrate the importance of various protocol features and show that it performs better than either the conservative or optimistic protocols in certain simulations.","PeriodicalId":162043,"journal":{"name":"1994 International Conference on Parallel Processing Vol. 3","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128193566","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Parallel Speculative Computation of Simulated Annealing","authors":"A. Sohn","doi":"10.1109/ICPP.1994.154","DOIUrl":"https://doi.org/10.1109/ICPP.1994.154","url":null,"abstract":"Simulated annealing is known to be highly sequential due to loop-carried dependencies. This report presents a new approach to parallel simulated annealing, called generalized speculative computation (GSC). We use an n-ary speculative tree and loop indices to execute n iterations in parallel on n processors while maintaining the same decision sequence as sequential simulated annealing. To verify the performance of GSC, we implement 100- to 500-city Traveling Salesman Problems on the AP1000 massively parallel multiprocessor. Execution results demonstrate that the GSC approach can indeed be an effective method for simulated annealing. We obtain over 20-fold speedup for the initial temperature of 0.1 and 11-fold speedup for the initial temperature of 10, all on 100 processors.","PeriodicalId":162043,"journal":{"name":"1994 International Conference on Parallel Processing Vol. 3","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1994-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126087091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fast Algorithms for Simulating the CRCW Shared-Memory Computer on Reconfigurable Meshes","authors":"Daw-Jong Shyu Daw-Jong Shyu, Biing-Feng Wang Biing-Feng Wang, Chuan-Yi Tang Chuan-Yi Tang","doi":"10.1109/ICPP.1994.112","DOIUrl":"https://doi.org/10.1109/ICPP.1994.112","url":null,"abstract":"In this paper, fast algorithms for simulating the CRCW shared-memory computer on reconfigurable meshes are proposed.","PeriodicalId":162043,"journal":{"name":"1994 International Conference on Parallel Processing Vol. 3","volume":"261 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115670924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}