Geometric semantic GP with linear scaling: Darwinian versus Lamarckian evolution

IF 1.7 3区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Giorgia Nadizar, Berfin Sakallioglu, Fraser Garrow, Sara Silva, Leonardo Vanneschi
{"title":"Geometric semantic GP with linear scaling: Darwinian versus Lamarckian evolution","authors":"Giorgia Nadizar, Berfin Sakallioglu, Fraser Garrow, Sara Silva, Leonardo Vanneschi","doi":"10.1007/s10710-024-09488-0","DOIUrl":null,"url":null,"abstract":"<p>Geometric Semantic Genetic Programming (GSGP) has shown notable success in symbolic regression with the introduction of Linear Scaling (LS). This achievement stems from the synergy of the geometric semantic genetic operators of GSGP with the scaling of the individuals for computing their fitness, which favours programs with a promising behaviour. However, the initial combination of GSGP and LS (GSGP-LS) underutilised the potential of LS, scaling individuals only for fitness evaluation, neglecting to incorporate improvements into their genetic material. In this paper we propose an advancement, GSGP with Lamarckian LS (GSGP-LLS), wherein we update the individuals in the population with their scaling coefficients in a Lamarckian fashion, i.e., by inheritance of acquired traits. We assess GSGP-LS and GSGP-LLS against standard GSGP for the task of symbolic regression on five hand-tailored benchmarks and six real-life problems. On the former ones, GSGP-LS and GSGP-LLS both consistently improve GSGP, though with no clear global superiority between them. On the real-world problems, instead, GSGP-LLS steadily outperforms GSGP-LS, achieving faster convergence and superior final performance. Notably, even in cases where LS induces overfitting on challenging problems, GSGP-LLS surpasses GSGP-LS, due to its slower and more localised optimisation steps.</p>","PeriodicalId":50424,"journal":{"name":"Genetic Programming and Evolvable Machines","volume":"9 1","pages":""},"PeriodicalIF":1.7000,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Genetic Programming and Evolvable Machines","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s10710-024-09488-0","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Geometric Semantic Genetic Programming (GSGP) has shown notable success in symbolic regression with the introduction of Linear Scaling (LS). This achievement stems from the synergy of the geometric semantic genetic operators of GSGP with the scaling of the individuals for computing their fitness, which favours programs with a promising behaviour. However, the initial combination of GSGP and LS (GSGP-LS) underutilised the potential of LS, scaling individuals only for fitness evaluation, neglecting to incorporate improvements into their genetic material. In this paper we propose an advancement, GSGP with Lamarckian LS (GSGP-LLS), wherein we update the individuals in the population with their scaling coefficients in a Lamarckian fashion, i.e., by inheritance of acquired traits. We assess GSGP-LS and GSGP-LLS against standard GSGP for the task of symbolic regression on five hand-tailored benchmarks and six real-life problems. On the former ones, GSGP-LS and GSGP-LLS both consistently improve GSGP, though with no clear global superiority between them. On the real-world problems, instead, GSGP-LLS steadily outperforms GSGP-LS, achieving faster convergence and superior final performance. Notably, even in cases where LS induces overfitting on challenging problems, GSGP-LLS surpasses GSGP-LS, due to its slower and more localised optimisation steps.

Abstract Image

线性缩放的几何语义 GP:达尔文进化论与拉马克进化论
几何语义遗传编程(GSGP)引入线性缩放(LS)后,在符号回归方面取得了显著的成功。这一成就源于 GSGP 的几何语义遗传算子与计算个体适合度的缩放的协同作用,这有利于具有良好行为的程序。然而,GSGP 和 LS 的最初组合(GSGP-LS)并没有充分利用 LS 的潜力,只是为了适配度评估而对个体进行缩放,忽略了将改进纳入其遗传物质中。在本文中,我们提出了一种改进方案,即带有拉马克 LS 的 GSGP(GSGP-LLS),通过拉马克方式(即后天性状的遗传)更新种群中个体的缩放系数。我们在五个手工定制的基准和六个实际问题上评估了 GSGP-LS 和 GSGP-LLS 与标准 GSGP 在符号回归任务上的对比。在前者中,GSGP-LS 和 GSGP-LLS 都持续改进了 GSGP,尽管它们之间没有明显的整体优势。相反,在实际问题上,GSGP-LLS 稳步超越 GSGP-LS,收敛速度更快,最终性能更优。值得注意的是,即使在 LS 引发过拟合的挑战性问题上,GSGP-LLS 也能超越 GSGP-LS,这是因为它的优化步骤更慢、更局部化。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Genetic Programming and Evolvable Machines
Genetic Programming and Evolvable Machines 工程技术-计算机:理论方法
CiteScore
5.90
自引率
3.80%
发文量
19
审稿时长
6 months
期刊介绍: A unique source reporting on methods for artificial evolution of programs and machines... Reports innovative and significant progress in automatic evolution of software and hardware. Features both theoretical and application papers. Covers hardware implementations, artificial life, molecular computing and emergent computation techniques. Examines such related topics as evolutionary algorithms with variable-size genomes, alternate methods of program induction, approaches to engineering systems development based on embryology, morphogenesis or other techniques inspired by adaptive natural systems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信