用于约束优化的具有自适应惩罚函数的协同进化算法

IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Vinícius Veloso de Melo, Alexandre Moreira Nascimento, Giovanni Iacca
{"title":"用于约束优化的具有自适应惩罚函数的协同进化算法","authors":"Vinícius Veloso de Melo, Alexandre Moreira Nascimento, Giovanni Iacca","doi":"10.1007/s00500-024-09896-5","DOIUrl":null,"url":null,"abstract":"<p>Several constrained optimization problems have been adequately solved over the years thanks to the advances in the area of metaheuristics. Nevertheless, the question as to which search logic performs better on constrained optimization often arises. In this paper, we present Dual Search Optimization (DSO), a co-evolutionary algorithm that includes an adaptive penalty function to handle constrained problems. Compared to other self-adaptive metaheuristics, one of the main advantages of DSO is that it is able auto-construct its own perturbation logics, i.e., the ways solutions are modified to create new ones during the optimization process. This is accomplished by co-evolving the solutions (encoded as vectors of integer/real values) and perturbation strategies (encoded as Genetic Programming trees), in order to adapt the search to the problem. In addition to that, the adaptive penalty function allows the algorithm to handle constraints very effectively, yet with a minor additional algorithmic overhead. We compare DSO with several algorithms from the state-of-the-art on two sets of problems, namely: (1) seven well-known constrained engineering design problems and (2) the CEC 2017 benchmark for constrained optimization. Our results show that DSO can achieve state-of-the-art performances, being capable to automatically adjust its behavior to the problem at hand.</p>","PeriodicalId":22039,"journal":{"name":"Soft Computing","volume":"86 1","pages":""},"PeriodicalIF":3.1000,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A co-evolutionary algorithm with adaptive penalty function for constrained optimization\",\"authors\":\"Vinícius Veloso de Melo, Alexandre Moreira Nascimento, Giovanni Iacca\",\"doi\":\"10.1007/s00500-024-09896-5\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Several constrained optimization problems have been adequately solved over the years thanks to the advances in the area of metaheuristics. Nevertheless, the question as to which search logic performs better on constrained optimization often arises. In this paper, we present Dual Search Optimization (DSO), a co-evolutionary algorithm that includes an adaptive penalty function to handle constrained problems. Compared to other self-adaptive metaheuristics, one of the main advantages of DSO is that it is able auto-construct its own perturbation logics, i.e., the ways solutions are modified to create new ones during the optimization process. This is accomplished by co-evolving the solutions (encoded as vectors of integer/real values) and perturbation strategies (encoded as Genetic Programming trees), in order to adapt the search to the problem. In addition to that, the adaptive penalty function allows the algorithm to handle constraints very effectively, yet with a minor additional algorithmic overhead. We compare DSO with several algorithms from the state-of-the-art on two sets of problems, namely: (1) seven well-known constrained engineering design problems and (2) the CEC 2017 benchmark for constrained optimization. Our results show that DSO can achieve state-of-the-art performances, being capable to automatically adjust its behavior to the problem at hand.</p>\",\"PeriodicalId\":22039,\"journal\":{\"name\":\"Soft Computing\",\"volume\":\"86 1\",\"pages\":\"\"},\"PeriodicalIF\":3.1000,\"publicationDate\":\"2024-07-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Soft Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.1007/s00500-024-09896-5\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s00500-024-09896-5","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

多年来,由于元启发式搜索技术的进步,一些约束优化问题得到了充分解决。然而,哪种搜索逻辑在约束优化中表现更好的问题经常出现。在本文中,我们介绍了双搜索优化(DSO),这是一种包含自适应惩罚函数的协同进化算法,用于处理约束问题。与其他自适应元启发式算法相比,DSO 的主要优势之一在于它能够自动构建自己的扰动逻辑,即在优化过程中修改解决方案以创建新解决方案的方式。这是通过共同演化解(编码为整数/实数向量)和扰动策略(编码为遗传编程树)来实现的,以便使搜索适应问题。此外,自适应惩罚函数允许算法非常有效地处理约束条件,但只需少量额外的算法开销。我们在两组问题上将 DSO 与最先进的几种算法进行了比较,这两组问题分别是:(1) 七个著名的约束工程设计问题;(2) CEC 2017 约束优化基准。我们的结果表明,DSO 可以实现最先进的性能,并能根据手头的问题自动调整其行为。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

A co-evolutionary algorithm with adaptive penalty function for constrained optimization

A co-evolutionary algorithm with adaptive penalty function for constrained optimization

Several constrained optimization problems have been adequately solved over the years thanks to the advances in the area of metaheuristics. Nevertheless, the question as to which search logic performs better on constrained optimization often arises. In this paper, we present Dual Search Optimization (DSO), a co-evolutionary algorithm that includes an adaptive penalty function to handle constrained problems. Compared to other self-adaptive metaheuristics, one of the main advantages of DSO is that it is able auto-construct its own perturbation logics, i.e., the ways solutions are modified to create new ones during the optimization process. This is accomplished by co-evolving the solutions (encoded as vectors of integer/real values) and perturbation strategies (encoded as Genetic Programming trees), in order to adapt the search to the problem. In addition to that, the adaptive penalty function allows the algorithm to handle constraints very effectively, yet with a minor additional algorithmic overhead. We compare DSO with several algorithms from the state-of-the-art on two sets of problems, namely: (1) seven well-known constrained engineering design problems and (2) the CEC 2017 benchmark for constrained optimization. Our results show that DSO can achieve state-of-the-art performances, being capable to automatically adjust its behavior to the problem at hand.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Soft Computing
Soft Computing 工程技术-计算机:跨学科应用
CiteScore
8.10
自引率
9.80%
发文量
927
审稿时长
7.3 months
期刊介绍: Soft Computing is dedicated to system solutions based on soft computing techniques. It provides rapid dissemination of important results in soft computing technologies, a fusion of research in evolutionary algorithms and genetic programming, neural science and neural net systems, fuzzy set theory and fuzzy systems, and chaos theory and chaotic systems. Soft Computing encourages the integration of soft computing techniques and tools into both everyday and advanced applications. By linking the ideas and techniques of soft computing with other disciplines, the journal serves as a unifying platform that fosters comparisons, extensions, and new applications. As a result, the journal is an international forum for all scientists and engineers engaged in research and development in this fast growing field.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信