{"title":"Incremental Refinement of Solutions for Dynamic Multi Objective Optimization Problems","authors":"C. E. Mariano-Romero, M.E.F. Morales","doi":"10.1109/MICAI.2007.47","DOIUrl":null,"url":null,"abstract":"MDQL is an algorithm, based on reinforcement learning, for solving multiple objective optimization problems, that has been tested on several applications with promising results. MDQL discretizes the decision variables into a set of states, each associated with actions to move agents to contiguous states. A group of agents explore this state space and are able to find Pareto sets applying a distributed reinforcement learning algorithm. The precision of the Pareto solutions depends on the chosen granularity of the states. A finer granularity on the states creates more precise solutions but at the expense of a larger search space, and consequently the need for more computational resources. In this paper, a very important improvement is introduced into the original MDQL algorithm to incrementally refined the Pareto solutions. The new algorithm, called IMDQL, starts with a coarse granularity to find an initial Pareto set. A vicinity for each of the Pareto solutions in refined and a new Pareto set is founded in this refined state space. This process continues until there is no more improvement within a small threshold value. It is shown that IMDQL not only improves the solutions found by MDQL, but also converges faster. MDQL has also been tested on the solutions of dynamic optimization problems. In this paper, it is also shown that the adaptation capabilities observed in MDQL can be improved with IMDQL. IMDQL was tested on the benchmark problems proposed by Jin. Performance evaluation was made using the Collective Mean Fitness metric proposed by Morrison. IMDQL was compared against an standard evolution strategy with the covariance matrix adaptation (CMA-ES) with very promising results.","PeriodicalId":296192,"journal":{"name":"2007 Sixth Mexican International Conference on Artificial Intelligence, Special Session (MICAI)","volume":"88 4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2007 Sixth Mexican International Conference on Artificial Intelligence, Special Session (MICAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MICAI.2007.47","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
MDQL is an algorithm, based on reinforcement learning, for solving multiple objective optimization problems, that has been tested on several applications with promising results. MDQL discretizes the decision variables into a set of states, each associated with actions to move agents to contiguous states. A group of agents explore this state space and are able to find Pareto sets applying a distributed reinforcement learning algorithm. The precision of the Pareto solutions depends on the chosen granularity of the states. A finer granularity on the states creates more precise solutions but at the expense of a larger search space, and consequently the need for more computational resources. In this paper, a very important improvement is introduced into the original MDQL algorithm to incrementally refined the Pareto solutions. The new algorithm, called IMDQL, starts with a coarse granularity to find an initial Pareto set. A vicinity for each of the Pareto solutions in refined and a new Pareto set is founded in this refined state space. This process continues until there is no more improvement within a small threshold value. It is shown that IMDQL not only improves the solutions found by MDQL, but also converges faster. MDQL has also been tested on the solutions of dynamic optimization problems. In this paper, it is also shown that the adaptation capabilities observed in MDQL can be improved with IMDQL. IMDQL was tested on the benchmark problems proposed by Jin. Performance evaluation was made using the Collective Mean Fitness metric proposed by Morrison. IMDQL was compared against an standard evolution strategy with the covariance matrix adaptation (CMA-ES) with very promising results.
MDQL是一种基于强化学习的算法,用于解决多目标优化问题,已经在几个应用程序中进行了测试,结果很有希望。MDQL将决策变量离散为一组状态,每个状态都与将代理移动到连续状态的操作相关联。一组智能体探索这个状态空间,并能够应用分布式强化学习算法找到帕累托集。帕累托解的精度取决于所选择的状态粒度。更细的状态粒度创建更精确的解决方案,但代价是更大的搜索空间,因此需要更多的计算资源。本文对原来的MDQL算法进行了一个非常重要的改进,以逐步改进Pareto解。这个名为IMDQL的新算法从粗粒度开始寻找初始帕累托集。在此精炼状态空间中建立了每个Pareto解的邻域,并建立了一个新的Pareto集。这个过程一直持续,直到在一个小的阈值内没有更多的改进。结果表明,IMDQL不仅改进了MDQL找到的解,而且收敛速度更快。MDQL还对动态优化问题的解决方案进行了测试。本文还表明,在MDQL中观察到的自适应能力可以通过IMDQL得到改进。IMDQL在Jin提出的基准问题上进行了测试。采用Morrison提出的集体平均适应度(Collective Mean Fitness)指标进行绩效评价。将IMDQL与具有协方差矩阵自适应(CMA-ES)的标准进化策略进行了比较,得到了很好的结果。