{"title":"MMDB partial reload","authors":"Le Gruenwald, Jing Huang","doi":"10.1006/jmca.1994.1008","DOIUrl":"https://doi.org/10.1006/jmca.1994.1008","url":null,"abstract":"<div><p>When a system failure occurs in a main memory database (MMDB) system, the primary copy of the database which resides permanently in a volatile memory is lost. Reload of the database from archive memory into main memory is needed to recover this lost copy. Several complete reload algorithms, which are used when main memory is large enough to store the entire database, have been designed for a parallel and multi-processor database environment. However, little research has been done for partial reload which is needed when only a part of the database can fit into main memory. This paper proposes three MMDB partial reload algorithms for a multi-processor main memory database system and presents their performance results with the Least Recently Used paging scheme being used to handle page faults. The simulation experiments show that reload based on frequency of data access yields the best transaction response time and system throughput.</p></div>","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"17 2","pages":"Pages 113-133"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1006/jmca.1994.1008","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71734059","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"MMDB partial reload","authors":"L. Gruenwald, Jing Huang","doi":"10.1006/JMCA.1994.1008","DOIUrl":"https://doi.org/10.1006/JMCA.1994.1008","url":null,"abstract":"Abstract When a system failure occurs in a main memory database (MMDB) system, the primary copy of the database which resides permanently in a volatile memory is lost. Reload of the database from archive memory into main memory is needed to recover this lost copy. Several complete reload algorithms, which are used when main memory is large enough to store the entire database, have been designed for a parallel and multi-processor database environment. However, little research has been done for partial reload which is needed when only a part of the database can fit into main memory. This paper proposes three MMDB partial reload algorithms for a multi-processor main memory database system and presents their performance results with the Least Recently Used paging scheme being used to handle page faults. The simulation experiments show that reload based on frequency of data access yields the best transaction response time and system throughput.","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"55 2","pages":"113-133"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91504504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Efficient barrier synchronization techniques and their applications in large-scale shared memory multiprocessors","authors":"K. Ghose, D. Cheng","doi":"10.1006/JMCA.1994.1012","DOIUrl":"https://doi.org/10.1006/JMCA.1994.1012","url":null,"abstract":"Abstract Shared memory multiprocessors offer a relatively simple programming model and are suitable for a wide variety of parallel applications. Unfortunately, shared memory multiprocessors are not as scalable as distributed memory multiprocessors owing to memory and switch contentions that can result in the formation of hot spots. Spinning on synchronization variables appears to be the main culprit behind the formation of hot spots, affecting system scalability adversely. The purpose of this paper is to address the issue of performing barrier synchronization efficiently in large-scale shared memory multiprocessors. We propose a very simple design for a hardware barrier synchronizer that has the characteristics of what one would call an ideal barrier synchronizer. In particular, the proposed barrier synchronizer allows fast barrier synchronization without injecting spin traffic to create hot spots and can be reused as soon as it has completed a barrier synchronization. We also show that by augmenting this barrier synchronizer with a few gates, it can be used to perform dynamic barrier synchronization, where neither the number, nor the exact identity of processors participating in the barrier is known a priori. We will also show that a low-latency barrier synchronizer can be used not only for high-speed barrier synchronization but also, very profitably, for implementing software combining (allowing distributed hot spot accessing), for data and producer-consumer type synchronization and for the implementation of a variety of other useful applications. A high-speed barrier synchronizer can also be used to implement highly concurrent data structures and will also allow a MIMD (Multiple Instruction streams, Multiple Data streams) system to be effectively operated in a SIMD (Single Instruction stream, Multiple Data streams)-style mode, giving rise to a number of potential advantages. We use simulations to confirm that our proposed synchronizers and their applications outperform the existing barrier synchronization schemes.","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"9 3","pages":"197-221"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91506503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Optimal cube-connected cube multicomputers","authors":"Jie Wu, Xian-He Sun","doi":"10.1006/jmca.1994.1009","DOIUrl":"https://doi.org/10.1006/jmca.1994.1009","url":null,"abstract":"<div><p>Many CFD (computational fluid dynamics) and other scientific applications can be partitioned into subproblems. However, in general, the partitioned subproblems are very large. They demand high-performance computing power themselves, and their solutions have to be combined at each time step. In this paper, the cube-connect cube (CCCube) architecture is studied. The CCCube architecture is an extended hypercube structure with each node represented as a cube. It requires fewer physical links between nodes than the hypercube, and provides the same communication support as the hypercube does on many applications. The reduced physical links can be used to enhance the bandwidth of the remanding links and, therefore, enhance the overall performance. The concept and the method to obtain optimal CCCubes, which are the CCCubes with a minimum number of links under a given total number of nodes, are proposed. The superiority of optimal CCCubes over standard hypercubes has also been shown in terms of the link usage in the embedding of a binomial tree. A useful computation structure based on a semi-binomial tree for divide-and-conquer type of parallel algorithms has been identified. We have shown that this structure can be implemented in optimal CCCubes without performance degradation compared with regular hypercubes. The result presented in this paper should provide a useful approach to design of scientific parallel computers.</p></div>","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"17 2","pages":"Pages 135-146"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1006/jmca.1994.1009","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71734056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multiprocessing in multiprotocol routers","authors":"D. Serpanos","doi":"10.1006/JMCA.1994.1007","DOIUrl":"https://doi.org/10.1006/JMCA.1994.1007","url":null,"abstract":"Abstract High-speed networks place strict requirements on the architecture of communication subsystems. One of the most significant problems in conventional subsystems is provision of high-speed protocol processing. The protocol processing problem is especially significant in the environment of multiprotocol routers where several routing protocols are supported. A multiprocessor architecture for high-speed processing in multiprotocol environments is presented and analyzed. It is shown that exploitation of vertical and horizontal parallelism in protocol stacks combined with parallelism in memory accesses and packet memory management significantly increases system performance. The presented architecture, used in realistic environments, meets the throughput requirements of high-speed network links offering throughput up to 100 Mbps.","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"1 1","pages":"99-112"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79904504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Special Issue on Parallel and Multiprocessor Systems","authors":"Nikolaos G. Bourbakis","doi":"10.1006/jmca.1994.1006","DOIUrl":"https://doi.org/10.1006/jmca.1994.1006","url":null,"abstract":"","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"17 2","pages":"Page 97"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1006/jmca.1994.1006","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71734002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A dynamically scheduled parallel DSP architecture for stream flow programming","authors":"Guoning Liao, Guang R. Gao, Vinod K. Agarwal","doi":"10.1006/jmca.1994.1011","DOIUrl":"https://doi.org/10.1006/jmca.1994.1011","url":null,"abstract":"<div><p>This paper presents a dynamically scheduled parallel DSP architecture for general purpose DSP computations. The architecture consists of multiple DSP processors and of one or more scheduling units. DSP applications are first captured by stream flow graphs, and then stream flow graphs are statically mapped onto a parallel architecture. The ordering and starting time of DSP tasks are determined by the scheduling unit(s) using a dynamic scheduling algorithm.</p><p>The main contributions of this paper are summarized as follows:</p><p>• A <em>scalable parallel DSP architecture</em>: The parallel DSP architecture proposed in this paper is scalable to meet signal processing requirements. For parallel DSP architectures with large configurations, the scheduling unit may become a performance bottleneck. A distributed scheduling mechanism is proposed to address this problem.</p><p>• A <em>mapping algorithm</em>: An algorithm is proposed to systematically map a stream flow graph onto a parallel DSP architecture.</p><p>• A <em>dynamic scheduling algorithm</em>: We propose a dynamic scheduling algorithm that will only schedule a node for execution when both input data and output storage space are available. Such scheduling algorithm will allow buffer sizes to be determined at compile time.</p><p>• A <em>simulation study</em>: Our simulation study reveals the relationships among the grain-size, the processor utilization, and the scheduling capability. We believe these relationships have significant impact on parallel computer architecture design involving dynamic scheduling.</p></div>","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"17 2","pages":"Pages 171-196"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1006/jmca.1994.1011","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71734058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Special Issue on Parallel and Multiprocessor Systems","authors":"N. Bourbakis","doi":"10.1006/jmca.1994.1006","DOIUrl":"https://doi.org/10.1006/jmca.1994.1006","url":null,"abstract":"","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"47 1","pages":"97"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78395461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A dynamically scheduled parallel DSP architecture for stream flow programming","authors":"Guoning Liao, G. Gao, V. Agarwal","doi":"10.1006/JMCA.1994.1011","DOIUrl":"https://doi.org/10.1006/JMCA.1994.1011","url":null,"abstract":"Abstract This paper presents a dynamically scheduled parallel DSP architecture for general purpose DSP computations. The architecture consists of multiple DSP processors and of one or more scheduling units. DSP applications are first captured by stream flow graphs, and then stream flow graphs are statically mapped onto a parallel architecture. The ordering and starting time of DSP tasks are determined by the scheduling unit(s) using a dynamic scheduling algorithm. The main contributions of this paper are summarized as follows: • A scalable parallel DSP architecture : The parallel DSP architecture proposed in this paper is scalable to meet signal processing requirements. For parallel DSP architectures with large configurations, the scheduling unit may become a performance bottleneck. A distributed scheduling mechanism is proposed to address this problem. • A mapping algorithm : An algorithm is proposed to systematically map a stream flow graph onto a parallel DSP architecture. • A dynamic scheduling algorithm : We propose a dynamic scheduling algorithm that will only schedule a node for execution when both input data and output storage space are available. Such scheduling algorithm will allow buffer sizes to be determined at compile time. • A simulation study : Our simulation study reveals the relationships among the grain-size, the processor utilization, and the scheduling capability. We believe these relationships have significant impact on parallel computer architecture design involving dynamic scheduling.","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"25 1","pages":"171-196"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88452131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A parallel hardware approach to implement network protocols","authors":"V. Catania, S. Cavalieri, A. Puliafito, L. Vita","doi":"10.1006/jmca.1994.1010","DOIUrl":"https://doi.org/10.1006/jmca.1994.1010","url":null,"abstract":"<div><p>The dramatic increase in the operating bandwidth of current communication networks requires a very high processing speed for communications protocols to offer users real broadband services.</p><p>Although there are different causes determining slowness in protocol processing, some implementation approaches for communication protocols seem to be the most limiting factor. This paper proposes a parallel architecture, based on VLSI components, to implement the OSI stack system in HSLANs and MANs. The proposed architecture is coupled with a model based on Petri nets which can characterize the implementation of each level according to the necessary requirements. We show that a hard protocol, as the data link layer, implemented according to the proposed architecture can effectively work at a data rate which is adequate for current networks.</p></div>","PeriodicalId":100806,"journal":{"name":"Journal of Microcomputer Applications","volume":"17 2","pages":"Pages 147-170"},"PeriodicalIF":0.0,"publicationDate":"1994-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1006/jmca.1994.1010","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"71734057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}