{"title":"Variance-reduced first-order methods for deterministically constrained stochastic nonconvex optimization with strong convergence guarantees","authors":"Zhaosong Lu, Sanyou Mei, Yifeng Xiao","doi":"arxiv-2409.09906","DOIUrl":null,"url":null,"abstract":"In this paper, we study a class of deterministically constrained stochastic\noptimization problems. Existing methods typically aim to find an\n$\\epsilon$-stochastic stationary point, where the expected violations of both\nthe constraints and first-order stationarity are within a prescribed accuracy\nof $\\epsilon$. However, in many practical applications, it is crucial that the\nconstraints be nearly satisfied with certainty, making such an\n$\\epsilon$-stochastic stationary point potentially undesirable due to the risk\nof significant constraint violations. To address this issue, we propose\nsingle-loop variance-reduced stochastic first-order methods, where the\nstochastic gradient of the stochastic component is computed using either a\ntruncated recursive momentum scheme or a truncated Polyak momentum scheme for\nvariance reduction, while the gradient of the deterministic component is\ncomputed exactly. Under the error bound condition with a parameter $\\theta \\geq\n1$ and other suitable assumptions, we establish that the proposed methods\nachieve a sample complexity and first-order operation complexity of $\\widetilde\nO(\\epsilon^{-\\max\\{4, 2\\theta\\}})$ for finding a stronger $\\epsilon$-stochastic\nstationary point, where the constraint violation is within $\\epsilon$ with\ncertainty, and the expected violation of first-order stationarity is within\n$\\epsilon$. To the best of our knowledge, this is the first work to develop\nmethods with provable complexity guarantees for finding an approximate\nstochastic stationary point of such problems that nearly satisfies all\nconstraints with certainty.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":"1 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09906","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, we study a class of deterministically constrained stochastic
optimization problems. Existing methods typically aim to find an
$\epsilon$-stochastic stationary point, where the expected violations of both
the constraints and first-order stationarity are within a prescribed accuracy
of $\epsilon$. However, in many practical applications, it is crucial that the
constraints be nearly satisfied with certainty, making such an
$\epsilon$-stochastic stationary point potentially undesirable due to the risk
of significant constraint violations. To address this issue, we propose
single-loop variance-reduced stochastic first-order methods, where the
stochastic gradient of the stochastic component is computed using either a
truncated recursive momentum scheme or a truncated Polyak momentum scheme for
variance reduction, while the gradient of the deterministic component is
computed exactly. Under the error bound condition with a parameter $\theta \geq
1$ and other suitable assumptions, we establish that the proposed methods
achieve a sample complexity and first-order operation complexity of $\widetilde
O(\epsilon^{-\max\{4, 2\theta\}})$ for finding a stronger $\epsilon$-stochastic
stationary point, where the constraint violation is within $\epsilon$ with
certainty, and the expected violation of first-order stationarity is within
$\epsilon$. To the best of our knowledge, this is the first work to develop
methods with provable complexity guarantees for finding an approximate
stochastic stationary point of such problems that nearly satisfies all
constraints with certainty.