Maximilian Reissmann, Yuan Fang, Andrew Ooi, Richard Sandberg
{"title":"通过语义反向传播限制遗传符号回归","authors":"Maximilian Reissmann, Yuan Fang, Andrew Ooi, Richard Sandberg","doi":"arxiv-2409.07369","DOIUrl":null,"url":null,"abstract":"Evolutionary symbolic regression approaches are powerful tools that can\napproximate an explicit mapping between input features and observation for\nvarious problems. However, ensuring that explored expressions maintain\nconsistency with domain-specific constraints remains a crucial challenge. While\nneural networks are able to employ additional information like conservation\nlaws to achieve more appropriate and robust approximations, the potential\nremains unrealized within genetic algorithms. This disparity is rooted in the\ninherent discrete randomness of recombining and mutating to generate new\nmapping expressions, making it challenging to maintain and preserve inferred\nconstraints or restrictions in the course of the exploration. To address this\nlimitation, we propose an approach centered on semantic backpropagation\nincorporated into the Gene Expression Programming (GEP), which integrates\ndomain-specific properties in a vector representation as corrective feedback\nduring the evolutionary process. By creating backward rules akin to algorithmic\ndifferentiation and leveraging pre-computed subsolutions, the mechanism allows\nthe enforcement of any constraint within an expression tree by determining the\nmisalignment and propagating desired changes back. To illustrate the\neffectiveness of constraining GEP through semantic backpropagation, we take the\nconstraint of physical dimension as an example. This framework is applied to\ndiscovering physical equations from the Feynman lectures. Results have shown\nnot only an increased likelihood of recovering the original equation but also\nnotable robustness in the presence of noisy data.","PeriodicalId":501286,"journal":{"name":"arXiv - MATH - Optimization and Control","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Constraining Genetic Symbolic Regression via Semantic Backpropagation\",\"authors\":\"Maximilian Reissmann, Yuan Fang, Andrew Ooi, Richard Sandberg\",\"doi\":\"arxiv-2409.07369\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Evolutionary symbolic regression approaches are powerful tools that can\\napproximate an explicit mapping between input features and observation for\\nvarious problems. However, ensuring that explored expressions maintain\\nconsistency with domain-specific constraints remains a crucial challenge. While\\nneural networks are able to employ additional information like conservation\\nlaws to achieve more appropriate and robust approximations, the potential\\nremains unrealized within genetic algorithms. This disparity is rooted in the\\ninherent discrete randomness of recombining and mutating to generate new\\nmapping expressions, making it challenging to maintain and preserve inferred\\nconstraints or restrictions in the course of the exploration. To address this\\nlimitation, we propose an approach centered on semantic backpropagation\\nincorporated into the Gene Expression Programming (GEP), which integrates\\ndomain-specific properties in a vector representation as corrective feedback\\nduring the evolutionary process. By creating backward rules akin to algorithmic\\ndifferentiation and leveraging pre-computed subsolutions, the mechanism allows\\nthe enforcement of any constraint within an expression tree by determining the\\nmisalignment and propagating desired changes back. To illustrate the\\neffectiveness of constraining GEP through semantic backpropagation, we take the\\nconstraint of physical dimension as an example. This framework is applied to\\ndiscovering physical equations from the Feynman lectures. Results have shown\\nnot only an increased likelihood of recovering the original equation but also\\nnotable robustness in the presence of noisy data.\",\"PeriodicalId\":501286,\"journal\":{\"name\":\"arXiv - MATH - Optimization and Control\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - MATH - Optimization and Control\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.07369\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Optimization and Control","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07369","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Constraining Genetic Symbolic Regression via Semantic Backpropagation
Evolutionary symbolic regression approaches are powerful tools that can
approximate an explicit mapping between input features and observation for
various problems. However, ensuring that explored expressions maintain
consistency with domain-specific constraints remains a crucial challenge. While
neural networks are able to employ additional information like conservation
laws to achieve more appropriate and robust approximations, the potential
remains unrealized within genetic algorithms. This disparity is rooted in the
inherent discrete randomness of recombining and mutating to generate new
mapping expressions, making it challenging to maintain and preserve inferred
constraints or restrictions in the course of the exploration. To address this
limitation, we propose an approach centered on semantic backpropagation
incorporated into the Gene Expression Programming (GEP), which integrates
domain-specific properties in a vector representation as corrective feedback
during the evolutionary process. By creating backward rules akin to algorithmic
differentiation and leveraging pre-computed subsolutions, the mechanism allows
the enforcement of any constraint within an expression tree by determining the
misalignment and propagating desired changes back. To illustrate the
effectiveness of constraining GEP through semantic backpropagation, we take the
constraint of physical dimension as an example. This framework is applied to
discovering physical equations from the Feynman lectures. Results have shown
not only an increased likelihood of recovering the original equation but also
notable robustness in the presence of noisy data.