{"title":"On Adaptive Stochastic Optimization for Streaming Data: A Newton's Method with O(dN) Operations","authors":"Antoine Godichon-BaggioniLPSM, Nicklas Werge","doi":"arxiv-2311.17753","DOIUrl":null,"url":null,"abstract":"Stochastic optimization methods encounter new challenges in the realm of\nstreaming, characterized by a continuous flow of large, high-dimensional data.\nWhile first-order methods, like stochastic gradient descent, are the natural\nchoice, they often struggle with ill-conditioned problems. In contrast,\nsecond-order methods, such as Newton's methods, offer a potential solution, but\ntheir computational demands render them impractical. This paper introduces\nadaptive stochastic optimization methods that bridge the gap between addressing\nill-conditioned problems while functioning in a streaming context. Notably, we\npresent an adaptive inversion-free Newton's method with a computational\ncomplexity matching that of first-order methods, $\\mathcal{O}(dN)$, where $d$\nrepresents the number of dimensions/features, and $N$ the number of data.\nTheoretical analysis confirms their asymptotic efficiency, and empirical\nevidence demonstrates their effectiveness, especially in scenarios involving\ncomplex covariance structures and challenging initializations. In particular,\nour adaptive Newton's methods outperform existing methods, while maintaining\nfavorable computational efficiency.","PeriodicalId":501330,"journal":{"name":"arXiv - MATH - Statistics Theory","volume":"89 5","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Statistics Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2311.17753","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Stochastic optimization methods encounter new challenges in the realm of
streaming, characterized by a continuous flow of large, high-dimensional data.
While first-order methods, like stochastic gradient descent, are the natural
choice, they often struggle with ill-conditioned problems. In contrast,
second-order methods, such as Newton's methods, offer a potential solution, but
their computational demands render them impractical. This paper introduces
adaptive stochastic optimization methods that bridge the gap between addressing
ill-conditioned problems while functioning in a streaming context. Notably, we
present an adaptive inversion-free Newton's method with a computational
complexity matching that of first-order methods, $\mathcal{O}(dN)$, where $d$
represents the number of dimensions/features, and $N$ the number of data.
Theoretical analysis confirms their asymptotic efficiency, and empirical
evidence demonstrates their effectiveness, especially in scenarios involving
complex covariance structures and challenging initializations. In particular,
our adaptive Newton's methods outperform existing methods, while maintaining
favorable computational efficiency.