{"title":"A Group Update Sparse Method Using Truncated Trust Region Strategy","authors":"Junxiang Li, Tao Dai, Feng Cheng, Jia-zhen Huo","doi":"10.1109/CSO.2011.31","DOIUrl":null,"url":null,"abstract":"We present a group update algorithm based on truncated trust region strategy for large-scale sparse unconstrained optimization. In large sparse optimization computing the whole Hessian matrix and solving exactly the Newton-like equations at each iteration can be considerably expensive. By the method the elements of the Hessian matrix are updated successively and periodically via groups during iterations and an inaccurate solution to the Newton-like equations is obtained by truncating the inner iteration under certain control rule. Besides, we allow that the current direction exceeds the trust region bound if it is a good descent direction satisfying some descent conditions. Some good convergence properties are kept and we contrast the computational behavior of our method with that of other algorithms. Our numerical tests show that the algorithm is promising and quite effective, and that its performance is comparable to or better than that of other algorithms available.","PeriodicalId":210815,"journal":{"name":"2011 Fourth International Joint Conference on Computational Sciences and Optimization","volume":"387 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 Fourth International Joint Conference on Computational Sciences and Optimization","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CSO.2011.31","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We present a group update algorithm based on truncated trust region strategy for large-scale sparse unconstrained optimization. In large sparse optimization computing the whole Hessian matrix and solving exactly the Newton-like equations at each iteration can be considerably expensive. By the method the elements of the Hessian matrix are updated successively and periodically via groups during iterations and an inaccurate solution to the Newton-like equations is obtained by truncating the inner iteration under certain control rule. Besides, we allow that the current direction exceeds the trust region bound if it is a good descent direction satisfying some descent conditions. Some good convergence properties are kept and we contrast the computational behavior of our method with that of other algorithms. Our numerical tests show that the algorithm is promising and quite effective, and that its performance is comparable to or better than that of other algorithms available.