Ruslan Abdulkadirov , Pavel Lyakhov , Maxim Bergerman , Nikolay Nagornov
{"title":"求解Navier-Stokes方程的快速差分梯度正负动量变压器算子网络","authors":"Ruslan Abdulkadirov , Pavel Lyakhov , Maxim Bergerman , Nikolay Nagornov","doi":"10.1016/j.chaos.2025.116964","DOIUrl":null,"url":null,"abstract":"<div><div>Modern machine learning approaches solve many problems in human activity. Recent models of neural networks find applications in solving equations of mathematical physics. Along with traditional numerical methods, physics-informed learning builds the approximate solution with a relatively small <span><math><msup><mrow><mi>L</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span> and mean squared errors. In this paper, we propose a fast difference-gradient positive–negative momentum optimizer that achieves a global minimum of the loss function with a convergence rate <span><math><mrow><mi>O</mi><mrow><mo>(</mo><mo>log</mo><mi>T</mi><mo>)</mo></mrow></mrow></math></span> and stability <span><math><mrow><mi>O</mi><mrow><mo>(</mo><mi>T</mi><mo>)</mo></mrow></mrow></math></span>. This optimization algorithm solves loss function minimization problems such as gradient discontinuity, vanishing gradient and local minimum traversal. Analysis on the set of test functions confirms the superiority of the proposed optimizer over state-of-the-art methods. Through the modified positive–negative moment estimation, the proposed optimizer gives more appropriate weight updates in physics-informed, deep operator, and transformer operator neural networks in Navier–Stokes equation solving. In particular, the proposed optimizer minimizes the loss function better than known analogs in solving <span><math><mrow><mn>2</mn><mi>d</mi></mrow></math></span>-Kovasznay, <span><math><mrow><mo>(</mo><mn>2</mn><mi>d</mi><mo>+</mo><mi>t</mi><mo>)</mo></mrow></math></span>-Taylor–Green, <span><math><mrow><mo>(</mo><mn>3</mn><mi>d</mi><mo>+</mo><mi>t</mi><mo>)</mo></mrow></math></span>-Beltrami, and <span><math><mrow><mo>(</mo><mn>2</mn><mi>d</mi><mo>+</mo><mi>t</mi><mo>)</mo></mrow></math></span> circular cylindrical flows. Fast difference gradient positive–negative moment estimation allows the physics-informed model to reduce the <span><math><msup><mrow><mi>L</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span> and mean squared errors solution by <span><math><mrow><mn>14</mn><mo>.</mo><mn>6</mn><mo>−</mo><mn>53</mn><mo>.</mo><mn>9</mn></mrow></math></span> percentage points. The proposed fast difference gradient positive–negative momentum can increase the approximate solutions of partial differential equations in physics-informed neural, deep operator, and transformer operator networks.</div></div>","PeriodicalId":9764,"journal":{"name":"Chaos Solitons & Fractals","volume":"200 ","pages":"Article 116964"},"PeriodicalIF":5.6000,"publicationDate":"2025-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Transformer operator network with fast difference gradient positive–negative momentum for solving Navier–Stokes equations\",\"authors\":\"Ruslan Abdulkadirov , Pavel Lyakhov , Maxim Bergerman , Nikolay Nagornov\",\"doi\":\"10.1016/j.chaos.2025.116964\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Modern machine learning approaches solve many problems in human activity. Recent models of neural networks find applications in solving equations of mathematical physics. Along with traditional numerical methods, physics-informed learning builds the approximate solution with a relatively small <span><math><msup><mrow><mi>L</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span> and mean squared errors. In this paper, we propose a fast difference-gradient positive–negative momentum optimizer that achieves a global minimum of the loss function with a convergence rate <span><math><mrow><mi>O</mi><mrow><mo>(</mo><mo>log</mo><mi>T</mi><mo>)</mo></mrow></mrow></math></span> and stability <span><math><mrow><mi>O</mi><mrow><mo>(</mo><mi>T</mi><mo>)</mo></mrow></mrow></math></span>. This optimization algorithm solves loss function minimization problems such as gradient discontinuity, vanishing gradient and local minimum traversal. Analysis on the set of test functions confirms the superiority of the proposed optimizer over state-of-the-art methods. Through the modified positive–negative moment estimation, the proposed optimizer gives more appropriate weight updates in physics-informed, deep operator, and transformer operator neural networks in Navier–Stokes equation solving. In particular, the proposed optimizer minimizes the loss function better than known analogs in solving <span><math><mrow><mn>2</mn><mi>d</mi></mrow></math></span>-Kovasznay, <span><math><mrow><mo>(</mo><mn>2</mn><mi>d</mi><mo>+</mo><mi>t</mi><mo>)</mo></mrow></math></span>-Taylor–Green, <span><math><mrow><mo>(</mo><mn>3</mn><mi>d</mi><mo>+</mo><mi>t</mi><mo>)</mo></mrow></math></span>-Beltrami, and <span><math><mrow><mo>(</mo><mn>2</mn><mi>d</mi><mo>+</mo><mi>t</mi><mo>)</mo></mrow></math></span> circular cylindrical flows. Fast difference gradient positive–negative moment estimation allows the physics-informed model to reduce the <span><math><msup><mrow><mi>L</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span> and mean squared errors solution by <span><math><mrow><mn>14</mn><mo>.</mo><mn>6</mn><mo>−</mo><mn>53</mn><mo>.</mo><mn>9</mn></mrow></math></span> percentage points. The proposed fast difference gradient positive–negative momentum can increase the approximate solutions of partial differential equations in physics-informed neural, deep operator, and transformer operator networks.</div></div>\",\"PeriodicalId\":9764,\"journal\":{\"name\":\"Chaos Solitons & Fractals\",\"volume\":\"200 \",\"pages\":\"Article 116964\"},\"PeriodicalIF\":5.6000,\"publicationDate\":\"2025-08-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Chaos Solitons & Fractals\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0960077925009774\",\"RegionNum\":1,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICS, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Chaos Solitons & Fractals","FirstCategoryId":"100","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0960077925009774","RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
Transformer operator network with fast difference gradient positive–negative momentum for solving Navier–Stokes equations
Modern machine learning approaches solve many problems in human activity. Recent models of neural networks find applications in solving equations of mathematical physics. Along with traditional numerical methods, physics-informed learning builds the approximate solution with a relatively small and mean squared errors. In this paper, we propose a fast difference-gradient positive–negative momentum optimizer that achieves a global minimum of the loss function with a convergence rate and stability . This optimization algorithm solves loss function minimization problems such as gradient discontinuity, vanishing gradient and local minimum traversal. Analysis on the set of test functions confirms the superiority of the proposed optimizer over state-of-the-art methods. Through the modified positive–negative moment estimation, the proposed optimizer gives more appropriate weight updates in physics-informed, deep operator, and transformer operator neural networks in Navier–Stokes equation solving. In particular, the proposed optimizer minimizes the loss function better than known analogs in solving -Kovasznay, -Taylor–Green, -Beltrami, and circular cylindrical flows. Fast difference gradient positive–negative moment estimation allows the physics-informed model to reduce the and mean squared errors solution by percentage points. The proposed fast difference gradient positive–negative momentum can increase the approximate solutions of partial differential equations in physics-informed neural, deep operator, and transformer operator networks.
期刊介绍:
Chaos, Solitons & Fractals strives to establish itself as a premier journal in the interdisciplinary realm of Nonlinear Science, Non-equilibrium, and Complex Phenomena. It welcomes submissions covering a broad spectrum of topics within this field, including dynamics, non-equilibrium processes in physics, chemistry, and geophysics, complex matter and networks, mathematical models, computational biology, applications to quantum and mesoscopic phenomena, fluctuations and random processes, self-organization, and social phenomena.