{"title":"基于词典的选择方法与符号回归问题的向下采样:概述与基准","authors":"Alina Geiger, Dominik Sobania, Franz Rothlauf","doi":"arxiv-2407.21632","DOIUrl":null,"url":null,"abstract":"In recent years, several new lexicase-based selection variants have emerged\ndue to the success of standard lexicase selection in various application\ndomains. For symbolic regression problems, variants that use an\nepsilon-threshold or batches of training cases, among others, have led to\nperformance improvements. Lately, especially variants that combine lexicase\nselection and down-sampling strategies have received a lot of attention. This\npaper evaluates random as well as informed down-sampling in combination with\nthe relevant lexicase-based selection methods on a wide range of symbolic\nregression problems. In contrast to most work, we not only compare the methods\nover a given evaluation budget, but also over a given time as time is usually\nlimited in practice. We find that for a given evaluation budget,\nepsilon-lexicase selection in combination with random or informed down-sampling\noutperforms all other methods. Only for a rather long running time of 24h, the\nbest performing method is tournament selection in combination with informed\ndown-sampling. If the given running time is very short, lexicase variants using\nbatches of training cases perform best.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"48 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Lexicase-based Selection Methods with Down-sampling for Symbolic Regression Problems: Overview and Benchmark\",\"authors\":\"Alina Geiger, Dominik Sobania, Franz Rothlauf\",\"doi\":\"arxiv-2407.21632\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In recent years, several new lexicase-based selection variants have emerged\\ndue to the success of standard lexicase selection in various application\\ndomains. For symbolic regression problems, variants that use an\\nepsilon-threshold or batches of training cases, among others, have led to\\nperformance improvements. Lately, especially variants that combine lexicase\\nselection and down-sampling strategies have received a lot of attention. This\\npaper evaluates random as well as informed down-sampling in combination with\\nthe relevant lexicase-based selection methods on a wide range of symbolic\\nregression problems. In contrast to most work, we not only compare the methods\\nover a given evaluation budget, but also over a given time as time is usually\\nlimited in practice. We find that for a given evaluation budget,\\nepsilon-lexicase selection in combination with random or informed down-sampling\\noutperforms all other methods. Only for a rather long running time of 24h, the\\nbest performing method is tournament selection in combination with informed\\ndown-sampling. If the given running time is very short, lexicase variants using\\nbatches of training cases perform best.\",\"PeriodicalId\":501347,\"journal\":{\"name\":\"arXiv - CS - Neural and Evolutionary Computing\",\"volume\":\"48 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-31\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Neural and Evolutionary Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2407.21632\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Neural and Evolutionary Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.21632","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Lexicase-based Selection Methods with Down-sampling for Symbolic Regression Problems: Overview and Benchmark
In recent years, several new lexicase-based selection variants have emerged
due to the success of standard lexicase selection in various application
domains. For symbolic regression problems, variants that use an
epsilon-threshold or batches of training cases, among others, have led to
performance improvements. Lately, especially variants that combine lexicase
selection and down-sampling strategies have received a lot of attention. This
paper evaluates random as well as informed down-sampling in combination with
the relevant lexicase-based selection methods on a wide range of symbolic
regression problems. In contrast to most work, we not only compare the methods
over a given evaluation budget, but also over a given time as time is usually
limited in practice. We find that for a given evaluation budget,
epsilon-lexicase selection in combination with random or informed down-sampling
outperforms all other methods. Only for a rather long running time of 24h, the
best performing method is tournament selection in combination with informed
down-sampling. If the given running time is very short, lexicase variants using
batches of training cases perform best.