{"title":"Efficient second-order optimization with predictions in differential games","authors":"Deliang Wei, Peng Chen, Fang Li, Xiangyun Zhang","doi":"10.1080/10556788.2023.2189715","DOIUrl":null,"url":null,"abstract":"A growing number of training methods for generative adversarial networks (GANs) are differential games. Different from convex optimization problems on single functions, gradient descent on multiple objectives may not converge to stable fixed points (SFPs). In order to improve learning dynamics in such games, many recently proposed methods utilize the second-order information of the game, such as the Hessian matrix. Unfortunately, these methods often suffer from the enormous computational cost of Hessian, which hinders their further applications. In this paper, we present efficient second-order optimization (ESO), in which only a part of Hessian is updated in each iteration, and the algorithm is derived. Furthermore, we give the local convergence of the method under reasonable assumptions. In order to further speed up the training process of GANs, we propose efficient second-order optimization with predictions (ESOP) using a novel accelerator. Basic experiments show that the proposed learning methods are faster than some state-of-art methods in GANs, while applicable to many other n-player differential games with local convergence guarantee.","PeriodicalId":124811,"journal":{"name":"Optimization Methods and Software","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Optimization Methods and Software","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/10556788.2023.2189715","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
A growing number of training methods for generative adversarial networks (GANs) are differential games. Different from convex optimization problems on single functions, gradient descent on multiple objectives may not converge to stable fixed points (SFPs). In order to improve learning dynamics in such games, many recently proposed methods utilize the second-order information of the game, such as the Hessian matrix. Unfortunately, these methods often suffer from the enormous computational cost of Hessian, which hinders their further applications. In this paper, we present efficient second-order optimization (ESO), in which only a part of Hessian is updated in each iteration, and the algorithm is derived. Furthermore, we give the local convergence of the method under reasonable assumptions. In order to further speed up the training process of GANs, we propose efficient second-order optimization with predictions (ESOP) using a novel accelerator. Basic experiments show that the proposed learning methods are faster than some state-of-art methods in GANs, while applicable to many other n-player differential games with local convergence guarantee.