{"title":"基于效用的反应网格强化学习","authors":"Julien Perez, C. Germain, B. Kégl, C. Loomis","doi":"10.1109/ICAC.2008.18","DOIUrl":null,"url":null,"abstract":"The main contribution of this paper is the presentation of a general scheduling framework for providing both QoS and fair-share in an autonomic fashion, based on 1) configurable utility functions and 2) RL as a model-free policy enactor. The main difference in our work is that we consider a multi-criteria optimization problem, including a fair-share objective. The comparison with a real and sophisticated scheduler shows that we could improve the most our RL scheme by accelerating the learning phase. More sophisticated interpolation (or regression) could speedup this phase. We plan to explore a hybrid scheme, where the RL is calibrated off-line by using the results of a real scheduler.","PeriodicalId":436716,"journal":{"name":"2008 International Conference on Autonomic Computing","volume":"55 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"24","resultStr":"{\"title\":\"Utility-Based Reinforcement Learning for Reactive Grids\",\"authors\":\"Julien Perez, C. Germain, B. Kégl, C. Loomis\",\"doi\":\"10.1109/ICAC.2008.18\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The main contribution of this paper is the presentation of a general scheduling framework for providing both QoS and fair-share in an autonomic fashion, based on 1) configurable utility functions and 2) RL as a model-free policy enactor. The main difference in our work is that we consider a multi-criteria optimization problem, including a fair-share objective. The comparison with a real and sophisticated scheduler shows that we could improve the most our RL scheme by accelerating the learning phase. More sophisticated interpolation (or regression) could speedup this phase. We plan to explore a hybrid scheme, where the RL is calibrated off-line by using the results of a real scheduler.\",\"PeriodicalId\":436716,\"journal\":{\"name\":\"2008 International Conference on Autonomic Computing\",\"volume\":\"55 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2008-06-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"24\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2008 International Conference on Autonomic Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICAC.2008.18\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 International Conference on Autonomic Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAC.2008.18","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Utility-Based Reinforcement Learning for Reactive Grids
The main contribution of this paper is the presentation of a general scheduling framework for providing both QoS and fair-share in an autonomic fashion, based on 1) configurable utility functions and 2) RL as a model-free policy enactor. The main difference in our work is that we consider a multi-criteria optimization problem, including a fair-share objective. The comparison with a real and sophisticated scheduler shows that we could improve the most our RL scheme by accelerating the learning phase. More sophisticated interpolation (or regression) could speedup this phase. We plan to explore a hybrid scheme, where the RL is calibrated off-line by using the results of a real scheduler.