{"title":"SPARC implementations: ASIC vs. custom design","authors":"M. Namjoo","doi":"10.1109/HICSS.1989.47139","DOIUrl":"https://doi.org/10.1109/HICSS.1989.47139","url":null,"abstract":"The first two implementations of the SPARC architecture, MB86900 and CY7601, were designed using high-speed CMOS technology with processor clock speed in the range of 16.6 to 33 MHz. In a system with a reasonable size external cache, these processors execute integer operations at a rate of approximately 1.5 clock cycles per instruction, resulting in a sustained performance in the range of 10 to 20 MIPS (millions of instructions per second). MB86900 design uses a single 20000-gate 1.3- mu m CMOS gate array and operates at a cycle time of 60 ns. CY7601 is a full custom chip designed using a 0.8- mu m CMOS process and operates at a cycle time of 30 ns. The basic features of these processors, their similarities and differences, and the tradeoffs used in their design. Design verification, test generation, and fault simulation are addressed.<<ETX>>","PeriodicalId":300182,"journal":{"name":"[1989] Proceedings of the Twenty-Second Annual Hawaii International Conference on System Sciences. Volume 1: Architecture Track","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125183375","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Guo‐zhen Yang, Yansong Chen, Shi-Hai Zhing, B. Dong, Dehuan Li
{"title":"A coherent system for performing an optical transform","authors":"Guo‐zhen Yang, Yansong Chen, Shi-Hai Zhing, B. Dong, Dehuan Li","doi":"10.1109/HICSS.1989.47188","DOIUrl":"https://doi.org/10.1109/HICSS.1989.47188","url":null,"abstract":"A coherent optical system for performing an arbitrary linear transform is described. The system consists of a holographic mask and two Fourier lenses. A set of equations for determining the amplitude-phase distribution of the mask is given, and the mask is generated by combination of a computer-generated hologram and optical holography. As an example, a Walsh-Hadamard transform of order 32 is realized.<<ETX>>","PeriodicalId":300182,"journal":{"name":"[1989] Proceedings of the Twenty-Second Annual Hawaii International Conference on System Sciences. Volume 1: Architecture Track","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126874967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Approaches to optical microprogramming","authors":"R. Forchheimer, V. Plavsic","doi":"10.1109/HICSS.1989.47191","DOIUrl":"https://doi.org/10.1109/HICSS.1989.47191","url":null,"abstract":"Approaches to the optical implementation of the stored-program concept are presented. The focus is on microprogramming, which is commonly used to organize data flow at the lowest hardware level. It is shown that the traditional implementation used in electronic computers maps easily into optics in many ways. A distributed addressing approach provides a simple solution, while a centralized addressing approach becomes more space-efficient as the number of instructions increases. A main contribution is the use of holographic storage for the control instructions. It is shown that full utilization of this storage technique can be achieved within a suggested interconnection structure.<<ETX>>","PeriodicalId":300182,"journal":{"name":"[1989] Proceedings of the Twenty-Second Annual Hawaii International Conference on System Sciences. Volume 1: Architecture Track","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125882972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Extending a Prolog architecture for high performance numeric computations","authors":"R. Yung, A. Despain, Y. Patt","doi":"10.1109/HICSS.1989.47181","DOIUrl":"https://doi.org/10.1109/HICSS.1989.47181","url":null,"abstract":"The Aquarius numeric processor (ANP) is an extended numeric instruction set architecture that is based on the Berkeley programmed logic machine (PLM) and supports integrated symbolic and numeric calculations. This extension expands the existing numeric data type to include 32- and 64-bit integers and single- and double-precision floating-point numbers conforming to the IEEE Standard P754. A class of data structure called numeric arrays has been added to represent matrices and arrays found in most scientific programming languages. Powerful numeric instructions are included to manipulate these novel data types. The authors describe the programming model and the architecture of the ANP. An experimental ANP is currently under construction using TTL (transistor-transistor logic) and ECL (emitter-coupled logic) parts. Simulated performance results indicate that the system will achieve about 10 MFLOPs (millions of floating-point operations) on the Prolog version of some Whetstone and Linpack benchmarks and close to 20 MFLOPS on some matrix operations (all in double precision).<<ETX>>","PeriodicalId":300182,"journal":{"name":"[1989] Proceedings of the Twenty-Second Annual Hawaii International Conference on System Sciences. Volume 1: Architecture Track","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128961935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A massive memory supercomputer","authors":"J. Rosenberg, D. Koch, J. Keedy","doi":"10.1109/HICSS.1989.47175","DOIUrl":"https://doi.org/10.1109/HICSS.1989.47175","url":null,"abstract":"An approach to supercomputing that is based on using a massive main memory (in the order of gigabytes) is investigated. Many of the problems currently solved on conventional supercomputers can equally be solved in similar time on such a machine, with a modest processor speed. The advantages of this approach in supporting database applications, VLSI applications and many other applications working on large volumes of date are examined. It is shown how the architecture of the MONADS-PC system, a capability-based computer developed by the authors, can be adapted to support such a large memory. The architectural design of a machine based on MONADS-PC is given, with special emphasis on the addressing and address translation issue.<<ETX>>","PeriodicalId":300182,"journal":{"name":"[1989] Proceedings of the Twenty-Second Annual Hawaii International Conference on System Sciences. Volume 1: Architecture Track","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114995335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A dedicated data flow architecture for hardware compilation","authors":"M. Naini","doi":"10.1109/HICSS.1989.47158","DOIUrl":"https://doi.org/10.1109/HICSS.1989.47158","url":null,"abstract":"A dedicated data-flow architecture that has been designed to be a part of a hardware compiler is described. This machine evaluates attribute grammars in a data-flow fashion by accepting their reverse-dependency graph, which is similar to a data-flow graph. The outputs and the results of these evaluations are sent to the other components for later use. The machine is believed to be the first dedicated data-flow architecture suggested for this purpose. It takes advantage of parallelism at two levels: first, the components of the machine are organized in a pipeline fashion and can run concurrently; second, the execution of the instructions is done in parallel as well.<<ETX>>","PeriodicalId":300182,"journal":{"name":"[1989] Proceedings of the Twenty-Second Annual Hawaii International Conference on System Sciences. Volume 1: Architecture Track","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1989-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131537969","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}