Baran Hashemi, Roderic G. Corominas, Alessandro Giacchetto
{"title":"变形金刚能做枚举几何吗?","authors":"Baran Hashemi, Roderic G. Corominas, Alessandro Giacchetto","doi":"arxiv-2408.14915","DOIUrl":null,"url":null,"abstract":"How can Transformers model and learn enumerative geometry? What is a robust\nprocedure for using Transformers in abductive knowledge discovery within a\nmathematician-machine collaboration? In this work, we introduce a new paradigm\nin computational enumerative geometry in analyzing the $\\psi$-class\nintersection numbers on the moduli space of curves. By formulating the\nenumerative problem as a continuous optimization task, we develop a\nTransformer-based model for computing $\\psi$-class intersection numbers based\non the underlying quantum Airy structure. For a finite range of genera, our\nmodel is capable of regressing intersection numbers that span an extremely wide\nrange of values, from $10^{-45}$ to $10^{45}$. To provide a proper inductive\nbias for capturing the recursive behavior of intersection numbers, we propose a\nnew activation function, Dynamic Range Activator (DRA). Moreover, given the\nsevere heteroscedasticity of $\\psi$-class intersections and the required\nprecision, we quantify the uncertainty of the predictions using Conformal\nPrediction with a dynamic sliding window that is aware of the number of marked\npoints. Next, we go beyond merely computing intersection numbers and explore\nthe enumerative \"world-model\" of the Transformers. Through a series of causal\ninference and correlational interpretability analyses, we demonstrate that\nTransformers are actually modeling Virasoro constraints in a purely data-driven\nmanner. Additionally, we provide evidence for the comprehension of several\nvalues appearing in the large genus asymptotic of $\\psi$-class intersection\nnumbers through abductive hypothesis testing.","PeriodicalId":501063,"journal":{"name":"arXiv - MATH - Algebraic Geometry","volume":"12 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Can Transformers Do Enumerative Geometry?\",\"authors\":\"Baran Hashemi, Roderic G. Corominas, Alessandro Giacchetto\",\"doi\":\"arxiv-2408.14915\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"How can Transformers model and learn enumerative geometry? What is a robust\\nprocedure for using Transformers in abductive knowledge discovery within a\\nmathematician-machine collaboration? In this work, we introduce a new paradigm\\nin computational enumerative geometry in analyzing the $\\\\psi$-class\\nintersection numbers on the moduli space of curves. By formulating the\\nenumerative problem as a continuous optimization task, we develop a\\nTransformer-based model for computing $\\\\psi$-class intersection numbers based\\non the underlying quantum Airy structure. For a finite range of genera, our\\nmodel is capable of regressing intersection numbers that span an extremely wide\\nrange of values, from $10^{-45}$ to $10^{45}$. To provide a proper inductive\\nbias for capturing the recursive behavior of intersection numbers, we propose a\\nnew activation function, Dynamic Range Activator (DRA). Moreover, given the\\nsevere heteroscedasticity of $\\\\psi$-class intersections and the required\\nprecision, we quantify the uncertainty of the predictions using Conformal\\nPrediction with a dynamic sliding window that is aware of the number of marked\\npoints. Next, we go beyond merely computing intersection numbers and explore\\nthe enumerative \\\"world-model\\\" of the Transformers. Through a series of causal\\ninference and correlational interpretability analyses, we demonstrate that\\nTransformers are actually modeling Virasoro constraints in a purely data-driven\\nmanner. Additionally, we provide evidence for the comprehension of several\\nvalues appearing in the large genus asymptotic of $\\\\psi$-class intersection\\nnumbers through abductive hypothesis testing.\",\"PeriodicalId\":501063,\"journal\":{\"name\":\"arXiv - MATH - Algebraic Geometry\",\"volume\":\"12 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - MATH - Algebraic Geometry\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.14915\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Algebraic Geometry","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.14915","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
How can Transformers model and learn enumerative geometry? What is a robust
procedure for using Transformers in abductive knowledge discovery within a
mathematician-machine collaboration? In this work, we introduce a new paradigm
in computational enumerative geometry in analyzing the $\psi$-class
intersection numbers on the moduli space of curves. By formulating the
enumerative problem as a continuous optimization task, we develop a
Transformer-based model for computing $\psi$-class intersection numbers based
on the underlying quantum Airy structure. For a finite range of genera, our
model is capable of regressing intersection numbers that span an extremely wide
range of values, from $10^{-45}$ to $10^{45}$. To provide a proper inductive
bias for capturing the recursive behavior of intersection numbers, we propose a
new activation function, Dynamic Range Activator (DRA). Moreover, given the
severe heteroscedasticity of $\psi$-class intersections and the required
precision, we quantify the uncertainty of the predictions using Conformal
Prediction with a dynamic sliding window that is aware of the number of marked
points. Next, we go beyond merely computing intersection numbers and explore
the enumerative "world-model" of the Transformers. Through a series of causal
inference and correlational interpretability analyses, we demonstrate that
Transformers are actually modeling Virasoro constraints in a purely data-driven
manner. Additionally, we provide evidence for the comprehension of several
values appearing in the large genus asymptotic of $\psi$-class intersection
numbers through abductive hypothesis testing.