{"title":"针对超标量和流水线处理器的编译器优化","authors":"Vishnu P. Bharadwaj, M. Rao","doi":"10.1109/DISCOVER.2016.7806224","DOIUrl":null,"url":null,"abstract":"The exploitation of parallelism at both the multiprocessor or multicore level and at the instruction level is the means to achieve high-performance. The compiler for VLIW and superscalar processors must expose sufficient parallelism to effectively utilize the parallel hardware. The amount of instruction level parallelism available to VLIW processors or superscalar processors can be limited. This will limit the performance of these processors to a certain extent. However, with compiler optimization techniques, its performance can be increased to greater extent. This evaluation shows that utilizing the existing resources of the processor with certain programmer constraints and an efficient scheduling of independent and dependent blocks of instructions, we can increase the performance of the processors. As compiler optimization interact with the micro-architecture in complex ways, certain programmer constraints can be added to reduce the complexity and help the compiler to structure the Assembly code in a manner which can be used for out-of-order execution of the code. This paper provides new methods and improvements for the structure of the Assembly code for execution on superscalar processors.","PeriodicalId":383554,"journal":{"name":"2016 IEEE Distributed Computing, VLSI, Electrical Circuits and Robotics (DISCOVER)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Compiler optimization for superscalar and pipelined processors\",\"authors\":\"Vishnu P. Bharadwaj, M. Rao\",\"doi\":\"10.1109/DISCOVER.2016.7806224\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The exploitation of parallelism at both the multiprocessor or multicore level and at the instruction level is the means to achieve high-performance. The compiler for VLIW and superscalar processors must expose sufficient parallelism to effectively utilize the parallel hardware. The amount of instruction level parallelism available to VLIW processors or superscalar processors can be limited. This will limit the performance of these processors to a certain extent. However, with compiler optimization techniques, its performance can be increased to greater extent. This evaluation shows that utilizing the existing resources of the processor with certain programmer constraints and an efficient scheduling of independent and dependent blocks of instructions, we can increase the performance of the processors. As compiler optimization interact with the micro-architecture in complex ways, certain programmer constraints can be added to reduce the complexity and help the compiler to structure the Assembly code in a manner which can be used for out-of-order execution of the code. This paper provides new methods and improvements for the structure of the Assembly code for execution on superscalar processors.\",\"PeriodicalId\":383554,\"journal\":{\"name\":\"2016 IEEE Distributed Computing, VLSI, Electrical Circuits and Robotics (DISCOVER)\",\"volume\":\"2 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2016 IEEE Distributed Computing, VLSI, Electrical Circuits and Robotics (DISCOVER)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/DISCOVER.2016.7806224\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2016 IEEE Distributed Computing, VLSI, Electrical Circuits and Robotics (DISCOVER)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DISCOVER.2016.7806224","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Compiler optimization for superscalar and pipelined processors
The exploitation of parallelism at both the multiprocessor or multicore level and at the instruction level is the means to achieve high-performance. The compiler for VLIW and superscalar processors must expose sufficient parallelism to effectively utilize the parallel hardware. The amount of instruction level parallelism available to VLIW processors or superscalar processors can be limited. This will limit the performance of these processors to a certain extent. However, with compiler optimization techniques, its performance can be increased to greater extent. This evaluation shows that utilizing the existing resources of the processor with certain programmer constraints and an efficient scheduling of independent and dependent blocks of instructions, we can increase the performance of the processors. As compiler optimization interact with the micro-architecture in complex ways, certain programmer constraints can be added to reduce the complexity and help the compiler to structure the Assembly code in a manner which can be used for out-of-order execution of the code. This paper provides new methods and improvements for the structure of the Assembly code for execution on superscalar processors.