Xu-Hui Zhou , Jiequn Han , Muhammad I. Zafar , Eric M. Wolf , Christopher R. Schrock , Christopher J. Roy , Heng Xiao
{"title":"Neural operator-based super-fidelity: A warm-start approach for accelerating steady-state simulations","authors":"Xu-Hui Zhou , Jiequn Han , Muhammad I. Zafar , Eric M. Wolf , Christopher R. Schrock , Christopher J. Roy , Heng Xiao","doi":"10.1016/j.jcp.2025.113871","DOIUrl":null,"url":null,"abstract":"<div><div>Neural networks have recently emerged as powerful tools for accelerated solving of partial differential equations (PDEs) in both academic and industrial settings. However, their use as standalone surrogate models raises concerns about reliability, as solution accuracy heavily depends on data quality, volume, and training algorithms. This concern is particularly pronounced in tasks that prioritize computational precision and deterministic outcomes. In response, this study introduces “super-fidelity”, a method that employs neural networks for initial warm-starts, significantly speeding up the solution of steady-state PDEs without compromising on accuracy. Drawing from super-resolution in computer vision, super-fidelity maps solutions from low-fidelity computational models to high-fidelity ones using a vector-cloud neural network with equivariance (VCNN-e)—a neural operator that preserves physical symmetries and adapts to different spatial discretizations. We evaluated the proposed method across scenarios with varying degrees of nonlinearity, including (1) two-dimensional laminar flows around elliptical cylinders at low Reynolds numbers, exhibiting monotonic convergence, (2) two-dimensional turbulent flows over airfoils at high Reynolds numbers, characterized by oscillatory convergence, and (3) practical three-dimensional turbulent flows over a wing. The results demonstrate that our neural operator-based initialization can accelerate convergence by at least a factor of two while maintaining the same level of accuracy, outperforming traditional initialization methods using uniform fields or potential flows. The approach's robustness and scalability are confirmed across different linear equation solvers and multi-process computing configurations. Additional investigations highlight its reduced dependence on high quality of training data, and real time savings across multiple simulations, even when including the neural-network model preparation time. Our study presents a promising strategy for accelerated solving of steady-state PDEs using neural operators, ensuring high accuracy in applications where precision is of utmost importance.</div></div>","PeriodicalId":352,"journal":{"name":"Journal of Computational Physics","volume":"529 ","pages":"Article 113871"},"PeriodicalIF":3.8000,"publicationDate":"2025-02-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Computational Physics","FirstCategoryId":"101","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0021999125001548","RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0
Abstract
Neural networks have recently emerged as powerful tools for accelerated solving of partial differential equations (PDEs) in both academic and industrial settings. However, their use as standalone surrogate models raises concerns about reliability, as solution accuracy heavily depends on data quality, volume, and training algorithms. This concern is particularly pronounced in tasks that prioritize computational precision and deterministic outcomes. In response, this study introduces “super-fidelity”, a method that employs neural networks for initial warm-starts, significantly speeding up the solution of steady-state PDEs without compromising on accuracy. Drawing from super-resolution in computer vision, super-fidelity maps solutions from low-fidelity computational models to high-fidelity ones using a vector-cloud neural network with equivariance (VCNN-e)—a neural operator that preserves physical symmetries and adapts to different spatial discretizations. We evaluated the proposed method across scenarios with varying degrees of nonlinearity, including (1) two-dimensional laminar flows around elliptical cylinders at low Reynolds numbers, exhibiting monotonic convergence, (2) two-dimensional turbulent flows over airfoils at high Reynolds numbers, characterized by oscillatory convergence, and (3) practical three-dimensional turbulent flows over a wing. The results demonstrate that our neural operator-based initialization can accelerate convergence by at least a factor of two while maintaining the same level of accuracy, outperforming traditional initialization methods using uniform fields or potential flows. The approach's robustness and scalability are confirmed across different linear equation solvers and multi-process computing configurations. Additional investigations highlight its reduced dependence on high quality of training data, and real time savings across multiple simulations, even when including the neural-network model preparation time. Our study presents a promising strategy for accelerated solving of steady-state PDEs using neural operators, ensuring high accuracy in applications where precision is of utmost importance.
期刊介绍:
Journal of Computational Physics thoroughly treats the computational aspects of physical problems, presenting techniques for the numerical solution of mathematical equations arising in all areas of physics. The journal seeks to emphasize methods that cross disciplinary boundaries.
The Journal of Computational Physics also publishes short notes of 4 pages or less (including figures, tables, and references but excluding title pages). Letters to the Editor commenting on articles already published in this Journal will also be considered. Neither notes nor letters should have an abstract.