{"title":"利用遗传算法进行调度","authors":"U. Fissgus","doi":"10.1109/ICDCS.2000.840983","DOIUrl":null,"url":null,"abstract":"Considers the scheduling of mixed task- and data-parallel modules comprising computation and communication operations. The program generation starts with a specification of the maximum degree of task- and data-parallelism of the method to be implemented. In several derivation steps, the degree of parallelism is adapted to a specific distributed memory machine. We present a scheduling derivation step based on the genetic algorithm paradigm. The scheduling takes not only decisions on the execution order (independent modules can be executed consecutively by all processors available or concurrently by independent groups of processors) but also on appropriate data distributions and task implementation versions. We demonstrate the efficiency of the algorithm by an example from numerical analysis.","PeriodicalId":284992,"journal":{"name":"Proceedings 20th IEEE International Conference on Distributed Computing Systems","volume":"207 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2000-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"19","resultStr":"{\"title\":\"Scheduling using genetic algorithms\",\"authors\":\"U. Fissgus\",\"doi\":\"10.1109/ICDCS.2000.840983\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Considers the scheduling of mixed task- and data-parallel modules comprising computation and communication operations. The program generation starts with a specification of the maximum degree of task- and data-parallelism of the method to be implemented. In several derivation steps, the degree of parallelism is adapted to a specific distributed memory machine. We present a scheduling derivation step based on the genetic algorithm paradigm. The scheduling takes not only decisions on the execution order (independent modules can be executed consecutively by all processors available or concurrently by independent groups of processors) but also on appropriate data distributions and task implementation versions. We demonstrate the efficiency of the algorithm by an example from numerical analysis.\",\"PeriodicalId\":284992,\"journal\":{\"name\":\"Proceedings 20th IEEE International Conference on Distributed Computing Systems\",\"volume\":\"207 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2000-04-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"19\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings 20th IEEE International Conference on Distributed Computing Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICDCS.2000.840983\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings 20th IEEE International Conference on Distributed Computing Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDCS.2000.840983","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Considers the scheduling of mixed task- and data-parallel modules comprising computation and communication operations. The program generation starts with a specification of the maximum degree of task- and data-parallelism of the method to be implemented. In several derivation steps, the degree of parallelism is adapted to a specific distributed memory machine. We present a scheduling derivation step based on the genetic algorithm paradigm. The scheduling takes not only decisions on the execution order (independent modules can be executed consecutively by all processors available or concurrently by independent groups of processors) but also on appropriate data distributions and task implementation versions. We demonstrate the efficiency of the algorithm by an example from numerical analysis.