{"title":"Parallel Kalman filtering on the Connection Machine","authors":"M.A. Palis, D.K. Krecker","doi":"10.1109/FMPC.1990.89438","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89438","url":null,"abstract":"A parallel algorithm for square-root Kalman filtering has been developed and implemented on the Connection Machine (CM). Performance measurements show that the CM filter runs in time linear in the state vector size. This represents a great improvement over serial implementations, which run in cubic time. A specific multiple-target-tracking application in which several targets are to be tracked simultaneously, each requiring one or more filters, is considered. A parallel algorithm that, for fixed-size filters, runs in constant time, independently of the number of filters simultaneously processed, has been developed.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122663744","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Data parallel computers and the FORALL statement","authors":"E. Albert, J. Lukas, G. Steele","doi":"10.1109/FMPC.1990.89489","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89489","url":null,"abstract":"The array constructs of Fortran 90 (formerly called Fortran 8*) map naturally onto SIMD (single-instruction-stream, multiple-data-stream) architectures that support a data parallel programming style, such as that of the Connection Machine computer system. The FORALL statement, an extension to Fortran 90 allowing for the expression of simultaneous execution of certain DO loop bodies, enhances this natural fit. A Fortran 90 compiler for data parallel machines is extended naturally to handle the FORALL statement. The data structures and algorithms used to effect this extension are described, and some examples of source code fragments and the target operations generated for them are presented.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115828384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A new parallel algorithm for the knapsack problem and its implementation on a hypercube","authors":"J. Lin, J. Storer","doi":"10.1109/FMPC.1990.89428","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89428","url":null,"abstract":"A new parallel algorithm is presented for the 0/1 knapsack problem. On a hypercube of n processors, the algorithm runs in time O(mc(log n)/n), where m is the number of objects and c is the knapsack size. The best previous known hypercube algorithm takes time O(mc/n+c log n+c/sup 2/). The new algorithm has been implemented on the Connection Machine and experimental results show that it performs very well for a wide range of problem sizes.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123040107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Parallel sorting of large arrays on the MasPar MP-1","authors":"J. Prins, J.A. Smith","doi":"10.1109/FMPC.1990.89439","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89439","url":null,"abstract":"The problem of sorting a collection of values on a mesh-connected, distributed-memory, SIMD (single-instruction-stream, multiple-data-stream) computer using variants of Batcher's bitonic sort algorithm is considered for the case in which the number of values exceeds the number of processors in the machine. In this setting the number of comparisons can be reduced asymptotically if the processors have addressing autonomy (locally indirect addressing), and communication costs can be reduced by judicious domain decomposition. The implementation of several related adaptations of bitonic sort on a MasPar MP-1 is reported. Performance is analyzed in relation to the virtualization ratio VPR. It is concluded that the most reasonable large-array sort for this machine will combine hypercube virtualization with the processor axes transposed dynamically within an xnet embedding.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121930462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Designing 3-D optical dilation multistage interconnection networks","authors":"J. Peir, K.-S. Huang","doi":"10.1109/FMPC.1990.89481","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89481","url":null,"abstract":"Three-dimensional multistage interconnection network designs based on a dilation structure are described. The dilation interconnection structure provides a space-invariant property among switching elements, which is suitable for constructing optical networks. Two different dilation topologies are discussed. One has the conventional rectangular shape, and the other uses the most space-compact hexagonal structure. Routing in the network is discussed.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121810013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Massive parallelism through program restructuring","authors":"M. Wolfe","doi":"10.1109/FMPC.1990.89491","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89491","url":null,"abstract":"A technique for mapping algorithms to massively parallel processors is described. It differs from previous work by focusing on explicit program restructuring, as opposed to manual or algebraic mapping. The method is flexible, and it allows nonlinear, as well as linear, mappings. Some restructuring transformations and how they would be used are described. A limitation of the approach is the restriction of skewing and rotating by unit factors only. The method benefits from previous work in program restructuring and systolic array synthesis and thus will be simple to implement.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1990-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132980048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Asymptotically efficient hypercube algorithms for computational geometry","authors":"P. MacKenzie, Q. Stout","doi":"10.1109/fmpc.1990.89429","DOIUrl":"https://doi.org/10.1109/fmpc.1990.89429","url":null,"abstract":"Hypercube algorithms that solve many fundamental computational geometry problems are presented. The algorithms use decomposition techniques, which enable them to outperform asymptotically the fastest previous algorithms for these problems. Previous algorithms all run in Theta (log/sup 2/n) time, even when using a sorting method that runs in o(log/sup 2/n) time. The new algorithms use a recently discovered o(log/sup 2/n) time sorting method to improve their asymptotic speed to o(log/sup 2/n). If sorting runs in Theta (Sort(n)) time, the algorithms for two-set dominance counting, 3-D maxima, closest pair, and all points nearest neighbors run in Theta (Sort(n)) log(log n) time, and the algorithms for triangulation and visibility from a point run in Theta (Sort(n)) time.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126114675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"How to use up processors","authors":"J. Hearne, D. Jusak","doi":"10.1109/FMPC.1990.89506","DOIUrl":"https://doi.org/10.1109/FMPC.1990.89506","url":null,"abstract":"Most parallelization aims for a decomposition in which the resulting units are data independent, each of the units contributes to the final output in the data flow, and synchronization is minimized. The paradigm of 'possible-worlds computing' aims to explore a model in which parallelization is achieved by disregarding the first two goals in order to preserve the third. The Tahiti programming language and its supporting run-time kernel Symphora, which are intended to explore the possible-worlds paradigm, are described.<<ETX>>","PeriodicalId":193332,"journal":{"name":"[1990 Proceedings] The Third Symposium on the Frontiers of Massively Parallel Computation","volume":"7 17","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132513153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}