Computational Optimization and Applications最新文献

筛选
英文 中文
T-semidefinite programming relaxation with third-order tensors for constrained polynomial optimization 用三阶张量对受约束多项式优化进行 T-半无限编程放松
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-06-02 DOI: 10.1007/s10589-024-00582-8
Hiroki Marumo, Sunyoung Kim, Makoto Yamashita
{"title":"T-semidefinite programming relaxation with third-order tensors for constrained polynomial optimization","authors":"Hiroki Marumo, Sunyoung Kim, Makoto Yamashita","doi":"10.1007/s10589-024-00582-8","DOIUrl":"https://doi.org/10.1007/s10589-024-00582-8","url":null,"abstract":"<p>We study T-semidefinite programming (SDP) relaxation for constrained polynomial optimization problems (POPs). T-SDP relaxation for unconstrained POPs was introduced by Zheng et al. (JGO 84:415–440, 2022). In this work, we propose a T-SDP relaxation for POPs with polynomial inequality constraints and show that the resulting T-SDP relaxation formulated with third-order tensors can be transformed into the standard SDP relaxation with block-diagonal structures. The convergence of the T-SDP relaxation to the optimal value of a given constrained POP is established under moderate assumptions as the relaxation level increases. Additionally, the feasibility and optimality of the T-SDP relaxation are discussed. Numerical results illustrate that the proposed T-SDP relaxation enhances numerical efficiency.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141195344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Delayed Weighted Gradient Method with simultaneous step-sizes for strongly convex optimization 采用同步步长的延迟加权梯度法进行强凸优化
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-05-31 DOI: 10.1007/s10589-024-00586-4
Hugo Lara, Rafael Aleixo, Harry Oviedo
{"title":"Delayed Weighted Gradient Method with simultaneous step-sizes for strongly convex optimization","authors":"Hugo Lara, Rafael Aleixo, Harry Oviedo","doi":"10.1007/s10589-024-00586-4","DOIUrl":"https://doi.org/10.1007/s10589-024-00586-4","url":null,"abstract":"<p>The Delayed Weighted Gradient Method (DWGM) is a two-step gradient algorithm that is efficient for the minimization of large scale strictly convex quadratic functions. It has orthogonality properties that make it to compete with the Conjugate Gradient (CG) method. Both methods calculate in sequence two step-sizes, CG minimizes the objective function and DWGM the gradient norm, alongside two search directions defined over first order current and previous iteration information. The objective of this work is to accelerate the recently developed extension of DWGM to nonquadratic strongly convex minimization problems. Our idea is to define the step-sizes of DWGM in a unique two dimensional convex quadratic optimization problem, calculating them simultaneously. Convergence of the resulting algorithm is analyzed. Comparative numerical experiments illustrate the effectiveness of our approach.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141195263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A non-monotone trust-region method with noisy oracles and additional sampling 具有噪声信标和额外采样的非单调信任区域方法
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-05-31 DOI: 10.1007/s10589-024-00580-w
Nataša Krejić, Nataša Krklec Jerinkić, Ángeles Martínez, Mahsa Yousefi
{"title":"A non-monotone trust-region method with noisy oracles and additional sampling","authors":"Nataša Krejić, Nataša Krklec Jerinkić, Ángeles Martínez, Mahsa Yousefi","doi":"10.1007/s10589-024-00580-w","DOIUrl":"https://doi.org/10.1007/s10589-024-00580-w","url":null,"abstract":"<p>In this work, we introduce a novel stochastic second-order method, within the framework of a non-monotone trust-region approach, for solving the unconstrained, nonlinear, and non-convex optimization problems arising in the training of deep neural networks. The proposed algorithm makes use of subsampling strategies that yield noisy approximations of the finite sum objective function and its gradient. We introduce an adaptive sample size strategy based on inexpensive additional sampling to control the resulting approximation error. Depending on the estimated progress of the algorithm, this can yield sample size scenarios ranging from mini-batch to full sample functions. We provide convergence analysis for all possible scenarios and show that the proposed method achieves almost sure convergence under standard assumptions for the trust-region framework. We report numerical experiments showing that the proposed algorithm outperforms its state-of-the-art counterpart in deep neural network training for image classification and regression tasks while requiring a significantly smaller number of gradient evaluations.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141195413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A boosted DC algorithm for non-differentiable DC components with non-monotone line search 采用非单调线搜索的无差别直流分量提升直流算法
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-05-11 DOI: 10.1007/s10589-024-00578-4
O. P. Ferreira, E. M. Santos, J. C. O. Souza
{"title":"A boosted DC algorithm for non-differentiable DC components with non-monotone line search","authors":"O. P. Ferreira, E. M. Santos, J. C. O. Souza","doi":"10.1007/s10589-024-00578-4","DOIUrl":"https://doi.org/10.1007/s10589-024-00578-4","url":null,"abstract":"<p>We introduce a new approach to apply the boosted difference of convex functions algorithm (BDCA) for solving non-convex and non-differentiable problems involving difference of two convex functions (DC functions). Supposing the first DC component differentiable and the second one possibly non-differentiable, the main idea of BDCA is to use the point computed by the subproblem of the DC algorithm (DCA) to define a descent direction of the objective from that point, and then a monotone line search starting from it is performed in order to find a new point which decreases the objective function when compared with the point generated by the subproblem of DCA. This procedure improves the performance of the DCA. However, if the first DC component is non-differentiable, then the direction computed by BDCA can be an ascent direction and a monotone line search cannot be performed. Our approach uses a non-monotone line search in the BDCA (nmBDCA) to enable a possible growth in the objective function values controlled by a parameter. Under suitable assumptions, we show that any cluster point of the sequence generated by the nmBDCA is a critical point of the problem under consideration and provides some iteration-complexity bounds. Furthermore, if the first DC component is differentiable, we present different iteration-complexity bounds and prove the full convergence of the sequence under the Kurdyka–Łojasiewicz property of the objective function. Some numerical experiments show that the nmBDCA outperforms the DCA, such as its monotone version.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140931603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An away-step Frank–Wolfe algorithm for constrained multiobjective optimization 用于约束性多目标优化的远离步骤弗兰克-沃尔夫算法
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-05-07 DOI: 10.1007/s10589-024-00577-5
Douglas S. Gonçalves, Max L. N. Gonçalves, Jefferson G. Melo
{"title":"An away-step Frank–Wolfe algorithm for constrained multiobjective optimization","authors":"Douglas S. Gonçalves, Max L. N. Gonçalves, Jefferson G. Melo","doi":"10.1007/s10589-024-00577-5","DOIUrl":"https://doi.org/10.1007/s10589-024-00577-5","url":null,"abstract":"<p>In this paper, we propose and analyze an <i>away-step</i> Frank–Wolfe algorithm designed for solving multiobjective optimization problems over polytopes. We prove that each limit point of the sequence generated by the algorithm is a weak Pareto optimal solution. Furthermore, under additional conditions, we show linear convergence of the whole sequence to a Pareto optimal solution. Numerical examples illustrate a promising performance of the proposed algorithm in problems where the multiobjective Frank–Wolfe convergence rate is only sublinear.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140887560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Shape optimization for interface identification in nonlocal models 非局部模型界面识别的形状优化
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-05-07 DOI: 10.1007/s10589-024-00575-7
Matthias Schuster, Christian Vollmann, Volker Schulz
{"title":"Shape optimization for interface identification in nonlocal models","authors":"Matthias Schuster, Christian Vollmann, Volker Schulz","doi":"10.1007/s10589-024-00575-7","DOIUrl":"https://doi.org/10.1007/s10589-024-00575-7","url":null,"abstract":"<p>Shape optimization methods have been proven useful for identifying interfaces in models governed by partial differential equations. Here we consider a class of shape optimization problems constrained by nonlocal equations which involve interface–dependent kernels. We derive a novel shape derivative associated to the nonlocal system model and solve the problem by established numerical techniques. The code for obtaining the results in this paper is published at (https://github.com/schustermatthias/nlshape).</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140931657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A hybrid inexact regularized Newton and negative curvature method 不精确正则化牛顿和负曲率混合方法
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-05-06 DOI: 10.1007/s10589-024-00576-6
Hong Zhu, Yunhai Xiao
{"title":"A hybrid inexact regularized Newton and negative curvature method","authors":"Hong Zhu, Yunhai Xiao","doi":"10.1007/s10589-024-00576-6","DOIUrl":"https://doi.org/10.1007/s10589-024-00576-6","url":null,"abstract":"<p>In this paper, we propose a hybrid inexact regularized Newton and negative curvature method for solving unconstrained nonconvex problems. The descent direction is chosen based on different conditions, either the negative curvature or the inexact regularized direction. In addition, to minimize computational costs while obtaining the negative curvature, we employ a dimensionality reduction strategy to verify if the Hessian matrix exhibits negative curvatures within a three-dimensional subspace. We show that the proposed method can achieve the best-known global iteration complexity if the Hessian of the objective function is Lipschitz continuous on a certain compact set. Two simplified methods for nonconvex and strongly convex problems are analyzed as specific instances of the proposed method. We show that under the local error bound assumption with respect to the gradient, the distance between iterations generated by our proposed method and the local solution set converges to <span>(0)</span> at a superlinear rate. Additionally, for strongly convex problems, the quadratic convergence rate can be achieved. Extensive numerical experiments show the effectiveness of the proposed method.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140887643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Chance-constrained programs with convex underlying functions: a bilevel convex optimization perspective 具有凸基础函数的机会受限程序:双层凸优化视角
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-04-27 DOI: 10.1007/s10589-024-00573-9
Yassine Laguel, Jérôme Malick, Wim van Ackooij
{"title":"Chance-constrained programs with convex underlying functions: a bilevel convex optimization perspective","authors":"Yassine Laguel, Jérôme Malick, Wim van Ackooij","doi":"10.1007/s10589-024-00573-9","DOIUrl":"https://doi.org/10.1007/s10589-024-00573-9","url":null,"abstract":"<p>Chance constraints are a valuable tool for the design of safe decisions in uncertain environments; they are used to model satisfaction of a constraint with a target probability. However, because of possible non-convexity and non-smoothness, optimizing over a chance constrained set is challenging. In this paper, we consider chance constrained programs where the objective function and the constraints are convex with respect to the decision parameter. We establish an exact reformulation of such a problem as a bilevel problem with a convex lower-level. Then we leverage this bilevel formulation to propose a tractable penalty approach, in the setting of finitely supported random variables. The penalized objective is a difference-of-convex function that we minimize with a suitable bundle algorithm. We release an easy-to-use open-source python toolbox implementing the approach, with a special emphasis on fast computational subroutines.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140809506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An inexactly accelerated algorithm for nonnegative tensor CP decomposition with the column unit constraints 带有列单位约束的非负张量 CP 分解的非精确加速算法
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-04-25 DOI: 10.1007/s10589-024-00574-8
Zihao Wang, Minru Bai
{"title":"An inexactly accelerated algorithm for nonnegative tensor CP decomposition with the column unit constraints","authors":"Zihao Wang, Minru Bai","doi":"10.1007/s10589-024-00574-8","DOIUrl":"https://doi.org/10.1007/s10589-024-00574-8","url":null,"abstract":"","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140653779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stochastic average model methods 随机平均模型方法
IF 2.2 2区 数学
Computational Optimization and Applications Pub Date : 2024-04-24 DOI: 10.1007/s10589-024-00563-x
Matt Menickelly, Stefan M. Wild
{"title":"Stochastic average model methods","authors":"Matt Menickelly, Stefan M. Wild","doi":"10.1007/s10589-024-00563-x","DOIUrl":"https://doi.org/10.1007/s10589-024-00563-x","url":null,"abstract":"<p>We consider the solution of finite-sum minimization problems, such as those appearing in nonlinear least-squares or general empirical risk minimization problems. We are motivated by problems in which the summand functions are computationally expensive and evaluating all summands on every iteration of an optimization method may be undesirable. We present the idea of stochastic average model (<span>SAM</span>) methods, inspired by stochastic average gradient methods. <span>SAM</span> methods sample component functions on each iteration of a trust-region method according to a discrete probability distribution on component functions; the distribution is designed to minimize an upper bound on the variance of the resulting stochastic model. We present promising numerical results concerning an implemented variant extending the derivative-free model-based trust-region solver <span>POUNDERS</span>, which we name <span>SAM-POUNDERS</span>.</p>","PeriodicalId":55227,"journal":{"name":"Computational Optimization and Applications","volume":null,"pages":null},"PeriodicalIF":2.2,"publicationDate":"2024-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140804672","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信