{"title":"A backward differential deep learning-based algorithm for solving high-dimensional nonlinear backward stochastic differential equations","authors":"Lorenc Kapllani, Long Teng","doi":"10.1093/imanum/draf022","DOIUrl":null,"url":null,"abstract":"In this work we propose a novel backward differential deep learning-based algorithm for solving high-dimensional nonlinear backward stochastic differential equations (BSDEs), where the deep neural network (DNN) models are trained, not only on the inputs and labels, but also on the differentials of the corresponding labels. This is motivated by the fact that differential deep learning can provide an efficient approximation of the labels and their derivatives with respect to inputs. The BSDEs are reformulated as differential deep learning problems by using Malliavin calculus. The Malliavin derivatives of the BSDE solution themselves satisfy another BSDE, resulting thus in a system of BSDEs. Such formulation requires the estimation of the solution, its gradient and the Hessian matrix, represented by the triple of processes $\\left (Y, Z, \\varGamma \\right ).$ All the integrals within this system are discretized by using the Euler–Maruyama method. Subsequently, DNNs are employed to approximate the triple of these unknown processes. The DNN parameters are backwardly optimized at each time step by minimizing a differential learning type loss function, which is defined as a weighted sum of the dynamics of the discretized BSDE system, with the first term providing the dynamics of the process $Y$ and the other the process $Z$. An error analysis is carried out to show the convergence of the proposed algorithm. Various numerical experiments of up to $50$ dimensions are provided to demonstrate the high efficiency. Both theoretically and numerically, it is demonstrated that our proposed scheme is more efficient in terms of computation time or accuracy compared with other contemporary deep learning-based methodologies.","PeriodicalId":56295,"journal":{"name":"IMA Journal of Numerical Analysis","volume":"39 1","pages":""},"PeriodicalIF":2.3000,"publicationDate":"2025-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IMA Journal of Numerical Analysis","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1093/imanum/draf022","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS, APPLIED","Score":null,"Total":0}
引用次数: 0
Abstract
In this work we propose a novel backward differential deep learning-based algorithm for solving high-dimensional nonlinear backward stochastic differential equations (BSDEs), where the deep neural network (DNN) models are trained, not only on the inputs and labels, but also on the differentials of the corresponding labels. This is motivated by the fact that differential deep learning can provide an efficient approximation of the labels and their derivatives with respect to inputs. The BSDEs are reformulated as differential deep learning problems by using Malliavin calculus. The Malliavin derivatives of the BSDE solution themselves satisfy another BSDE, resulting thus in a system of BSDEs. Such formulation requires the estimation of the solution, its gradient and the Hessian matrix, represented by the triple of processes $\left (Y, Z, \varGamma \right ).$ All the integrals within this system are discretized by using the Euler–Maruyama method. Subsequently, DNNs are employed to approximate the triple of these unknown processes. The DNN parameters are backwardly optimized at each time step by minimizing a differential learning type loss function, which is defined as a weighted sum of the dynamics of the discretized BSDE system, with the first term providing the dynamics of the process $Y$ and the other the process $Z$. An error analysis is carried out to show the convergence of the proposed algorithm. Various numerical experiments of up to $50$ dimensions are provided to demonstrate the high efficiency. Both theoretically and numerically, it is demonstrated that our proposed scheme is more efficient in terms of computation time or accuracy compared with other contemporary deep learning-based methodologies.
期刊介绍:
The IMA Journal of Numerical Analysis (IMAJNA) publishes original contributions to all fields of numerical analysis; articles will be accepted which treat the theory, development or use of practical algorithms and interactions between these aspects. Occasional survey articles are also published.