{"title":"动态调度基准:设计、实现和性能评价","authors":"B. Hamidzadeh, Ngan Kee","doi":"10.1109/TAI.1999.809806","DOIUrl":null,"url":null,"abstract":"Puzzles have traditionally been used as popular benchmarks for evaluating different problem solving strategies. Many of the game benchmarks are suitable for evaluating static scheduling techniques. In such benchmarks, the scheduling phase and the execution phase (i.e. when the schedule is executed to play the game) are disjoint. The scheduling technique can be executed to compute a complete schedule prior to the execution of any moves to play the game. Due to recent interest in on-line problem solving techniques, there is a need for benchmarks which can evaluate the performance trade-offs of dynamic scheduling techniques. Many modern video and computer games can be suitable candidates for dynamic scheduling benchmarks, since they require on-line problem solving. These benchmarks and their system testbeds should be chosen and implemented such that they can accurately reveal important performance trade-offs of dynamic scheduling techniques. In this paper, we introduce a dynamic scheduling benchmark and its system testbed. This benchmark is based on an extended version of the Tetris computer game. The rules and semantics of the game were modified to lend themselves well to evaluation of discrete problem solving and optimization techniques. The system testbed is implemented in a distributed and asynchronous fashion, on a network of workstations, to reveal performance trade-offs between scheduling time, schedule quality, and problem constraints.","PeriodicalId":194023,"journal":{"name":"Proceedings 11th International Conference on Tools with Artificial Intelligence","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1999-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A dynamic scheduling benchmark: design, implementation and performance evaluation\",\"authors\":\"B. Hamidzadeh, Ngan Kee\",\"doi\":\"10.1109/TAI.1999.809806\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Puzzles have traditionally been used as popular benchmarks for evaluating different problem solving strategies. Many of the game benchmarks are suitable for evaluating static scheduling techniques. In such benchmarks, the scheduling phase and the execution phase (i.e. when the schedule is executed to play the game) are disjoint. The scheduling technique can be executed to compute a complete schedule prior to the execution of any moves to play the game. Due to recent interest in on-line problem solving techniques, there is a need for benchmarks which can evaluate the performance trade-offs of dynamic scheduling techniques. Many modern video and computer games can be suitable candidates for dynamic scheduling benchmarks, since they require on-line problem solving. These benchmarks and their system testbeds should be chosen and implemented such that they can accurately reveal important performance trade-offs of dynamic scheduling techniques. In this paper, we introduce a dynamic scheduling benchmark and its system testbed. This benchmark is based on an extended version of the Tetris computer game. The rules and semantics of the game were modified to lend themselves well to evaluation of discrete problem solving and optimization techniques. The system testbed is implemented in a distributed and asynchronous fashion, on a network of workstations, to reveal performance trade-offs between scheduling time, schedule quality, and problem constraints.\",\"PeriodicalId\":194023,\"journal\":{\"name\":\"Proceedings 11th International Conference on Tools with Artificial Intelligence\",\"volume\":\"23 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1999-11-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings 11th International Conference on Tools with Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/TAI.1999.809806\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings 11th International Conference on Tools with Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/TAI.1999.809806","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A dynamic scheduling benchmark: design, implementation and performance evaluation
Puzzles have traditionally been used as popular benchmarks for evaluating different problem solving strategies. Many of the game benchmarks are suitable for evaluating static scheduling techniques. In such benchmarks, the scheduling phase and the execution phase (i.e. when the schedule is executed to play the game) are disjoint. The scheduling technique can be executed to compute a complete schedule prior to the execution of any moves to play the game. Due to recent interest in on-line problem solving techniques, there is a need for benchmarks which can evaluate the performance trade-offs of dynamic scheduling techniques. Many modern video and computer games can be suitable candidates for dynamic scheduling benchmarks, since they require on-line problem solving. These benchmarks and their system testbeds should be chosen and implemented such that they can accurately reveal important performance trade-offs of dynamic scheduling techniques. In this paper, we introduce a dynamic scheduling benchmark and its system testbed. This benchmark is based on an extended version of the Tetris computer game. The rules and semantics of the game were modified to lend themselves well to evaluation of discrete problem solving and optimization techniques. The system testbed is implemented in a distributed and asynchronous fashion, on a network of workstations, to reveal performance trade-offs between scheduling time, schedule quality, and problem constraints.