J. Z. Salazar, D. Hadka, P. Reed, H. Seada, K. Deb
{"title":"评估多目标进化算法:一个真实世界的基准框架","authors":"J. Z. Salazar, D. Hadka, P. Reed, H. Seada, K. Deb","doi":"10.36334/modsim.2023.salazar","DOIUrl":null,"url":null,"abstract":": Multiobjective evolutionary algorithms (MOEAs) have shown significant progress in addressing well-defined test problems, but their effectiveness in real-world applications remains uncertain. To bridge this gap, we provide a comprehensive benchmarking framework designed to rigorously evaluate state-of-the-art MOEAs in real-world scenarios. Our framework comprises a suite of statistical evaluation metrics, for a collection of diverse real-world benchmark applications representing various mathematical problem difficulties. In this study, we carefully selected four benchmark applications with 3 to 10 objectives, capturing challenging characteristics such as stochastic objectives, severe constraints, strong nonlinearity, and complex Pareto frontiers. We evaluated the performance of five popular MOEAs, including NSGA-II, NSGA-III, RVEA, MOEA/D, and the Borg MOEA, using our benchmarking framework. Multiobjective evolutionary algorithms (MOEAs) have shown significant progress in addressing well-defined test problems, but their effectiveness in real-world applications remains uncertain. To bridge this gap, we provide a comprehensive benchmarking framework designed to rigorously evaluate state-of-the-art MOEAs in real-world scenarios. Our framework comprises a suite of statistical evaluation metrics, for a collection of diverse real-world benchmark applications representing various mathematical problem difficulties. The results revealed distinct differences in the performance of the evaluated MOEAs across the real-world applications. Surprisingly, MOEAs that excelled on standard test problems struggled when confronted with the complexities inherent in real-world applications. These findings underscore the need to enhance the adaptability and ease-of-use of MOEAs, considering the often ill-defined nature of real-world problem solving. Furthermore, our study provides insights into successful algorithmic design choices for MOEAs. Optimal selection strategies and archive mechanisms are crucial to prevent deterioration, maintain diversity, and provide adequate selection pressure throughout the optimization process. Additionally, the choice of stable and flexible operators plays a vital role in reliably driving the search towards the Pareto front. Recent advancements in hyper-heuristics and multi-operator MOEAs offer promising automated approaches for tackling these challenges. We found that epsilon non-dominated sorting effectively maintains diversity and selection pressure for problems with up to ten objectives when the entire Pareto front is desired. Moreover, auto-adaptive search operators demonstrate their efficacy in adapting to the search landscape of diverse real-world applications. However, the performance of reference point/vector methods deteriorated at higher dimensions, indicating the need for further investigation. Our study highlights the inadequacy of existing test benchmarks in differentiating MOEAs based on real-world performance. While considerable efforts have focused on optimizing algorithms for test problems, the subpar performance of MOEAs in real-world settings persists. The benchmarking framework and results presented here aim to foster collaborative efforts, encouraging the development of a diverse suite of real-world benchmarking problems. The flexibility of our framework also allows for exploration of new or hybrid algorithm architectures in benchmarking studies.","PeriodicalId":390064,"journal":{"name":"MODSIM2023, 25th International Congress on Modelling and Simulation.","volume":"68 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Evaluating multiobjective evolutionary algorithms: A real-world benchmarking framework\",\"authors\":\"J. Z. Salazar, D. Hadka, P. Reed, H. Seada, K. Deb\",\"doi\":\"10.36334/modsim.2023.salazar\",\"DOIUrl\":null,\"url\":null,\"abstract\":\": Multiobjective evolutionary algorithms (MOEAs) have shown significant progress in addressing well-defined test problems, but their effectiveness in real-world applications remains uncertain. To bridge this gap, we provide a comprehensive benchmarking framework designed to rigorously evaluate state-of-the-art MOEAs in real-world scenarios. Our framework comprises a suite of statistical evaluation metrics, for a collection of diverse real-world benchmark applications representing various mathematical problem difficulties. In this study, we carefully selected four benchmark applications with 3 to 10 objectives, capturing challenging characteristics such as stochastic objectives, severe constraints, strong nonlinearity, and complex Pareto frontiers. We evaluated the performance of five popular MOEAs, including NSGA-II, NSGA-III, RVEA, MOEA/D, and the Borg MOEA, using our benchmarking framework. Multiobjective evolutionary algorithms (MOEAs) have shown significant progress in addressing well-defined test problems, but their effectiveness in real-world applications remains uncertain. To bridge this gap, we provide a comprehensive benchmarking framework designed to rigorously evaluate state-of-the-art MOEAs in real-world scenarios. Our framework comprises a suite of statistical evaluation metrics, for a collection of diverse real-world benchmark applications representing various mathematical problem difficulties. The results revealed distinct differences in the performance of the evaluated MOEAs across the real-world applications. Surprisingly, MOEAs that excelled on standard test problems struggled when confronted with the complexities inherent in real-world applications. These findings underscore the need to enhance the adaptability and ease-of-use of MOEAs, considering the often ill-defined nature of real-world problem solving. Furthermore, our study provides insights into successful algorithmic design choices for MOEAs. Optimal selection strategies and archive mechanisms are crucial to prevent deterioration, maintain diversity, and provide adequate selection pressure throughout the optimization process. Additionally, the choice of stable and flexible operators plays a vital role in reliably driving the search towards the Pareto front. Recent advancements in hyper-heuristics and multi-operator MOEAs offer promising automated approaches for tackling these challenges. We found that epsilon non-dominated sorting effectively maintains diversity and selection pressure for problems with up to ten objectives when the entire Pareto front is desired. Moreover, auto-adaptive search operators demonstrate their efficacy in adapting to the search landscape of diverse real-world applications. However, the performance of reference point/vector methods deteriorated at higher dimensions, indicating the need for further investigation. Our study highlights the inadequacy of existing test benchmarks in differentiating MOEAs based on real-world performance. While considerable efforts have focused on optimizing algorithms for test problems, the subpar performance of MOEAs in real-world settings persists. The benchmarking framework and results presented here aim to foster collaborative efforts, encouraging the development of a diverse suite of real-world benchmarking problems. The flexibility of our framework also allows for exploration of new or hybrid algorithm architectures in benchmarking studies.\",\"PeriodicalId\":390064,\"journal\":{\"name\":\"MODSIM2023, 25th International Congress on Modelling and Simulation.\",\"volume\":\"68 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"MODSIM2023, 25th International Congress on Modelling and Simulation.\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.36334/modsim.2023.salazar\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"MODSIM2023, 25th International Congress on Modelling and Simulation.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.36334/modsim.2023.salazar","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Evaluating multiobjective evolutionary algorithms: A real-world benchmarking framework
: Multiobjective evolutionary algorithms (MOEAs) have shown significant progress in addressing well-defined test problems, but their effectiveness in real-world applications remains uncertain. To bridge this gap, we provide a comprehensive benchmarking framework designed to rigorously evaluate state-of-the-art MOEAs in real-world scenarios. Our framework comprises a suite of statistical evaluation metrics, for a collection of diverse real-world benchmark applications representing various mathematical problem difficulties. In this study, we carefully selected four benchmark applications with 3 to 10 objectives, capturing challenging characteristics such as stochastic objectives, severe constraints, strong nonlinearity, and complex Pareto frontiers. We evaluated the performance of five popular MOEAs, including NSGA-II, NSGA-III, RVEA, MOEA/D, and the Borg MOEA, using our benchmarking framework. Multiobjective evolutionary algorithms (MOEAs) have shown significant progress in addressing well-defined test problems, but their effectiveness in real-world applications remains uncertain. To bridge this gap, we provide a comprehensive benchmarking framework designed to rigorously evaluate state-of-the-art MOEAs in real-world scenarios. Our framework comprises a suite of statistical evaluation metrics, for a collection of diverse real-world benchmark applications representing various mathematical problem difficulties. The results revealed distinct differences in the performance of the evaluated MOEAs across the real-world applications. Surprisingly, MOEAs that excelled on standard test problems struggled when confronted with the complexities inherent in real-world applications. These findings underscore the need to enhance the adaptability and ease-of-use of MOEAs, considering the often ill-defined nature of real-world problem solving. Furthermore, our study provides insights into successful algorithmic design choices for MOEAs. Optimal selection strategies and archive mechanisms are crucial to prevent deterioration, maintain diversity, and provide adequate selection pressure throughout the optimization process. Additionally, the choice of stable and flexible operators plays a vital role in reliably driving the search towards the Pareto front. Recent advancements in hyper-heuristics and multi-operator MOEAs offer promising automated approaches for tackling these challenges. We found that epsilon non-dominated sorting effectively maintains diversity and selection pressure for problems with up to ten objectives when the entire Pareto front is desired. Moreover, auto-adaptive search operators demonstrate their efficacy in adapting to the search landscape of diverse real-world applications. However, the performance of reference point/vector methods deteriorated at higher dimensions, indicating the need for further investigation. Our study highlights the inadequacy of existing test benchmarks in differentiating MOEAs based on real-world performance. While considerable efforts have focused on optimizing algorithms for test problems, the subpar performance of MOEAs in real-world settings persists. The benchmarking framework and results presented here aim to foster collaborative efforts, encouraging the development of a diverse suite of real-world benchmarking problems. The flexibility of our framework also allows for exploration of new or hybrid algorithm architectures in benchmarking studies.