{"title":"Deep learning from strongly mixing observations: Sparse-penalized regularization and minimax optimality","authors":"William Kengne , Modou Wade","doi":"10.1016/j.jco.2025.101978","DOIUrl":null,"url":null,"abstract":"<div><div>This paper considers deep learning from strongly mixing observations and performs a sparse-penalized regularization for deep neural networks (DNN) predictors. In a general framework that includes regression and classification, oracle inequalities for the expected excess risk are established, and upper bounds on the class of Hölder smooth functions and composition structured Hölder functions are provided. For nonparametric autoregression with the Gaussian and Laplace errors, and the Huber loss function, it is shown that the sparse-penalized DNN estimator proposed is optimal (up to a logarithmic factor) in the minimax sense. Based on the lower bound established in Alquier and Kengne (2024), we show that the proposed DNN estimator for the classification task with the logistic loss on strongly mixing observations achieves (up to a logarithmic factor), the minimax optimal convergence rate.</div></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"92 ","pages":"Article 101978"},"PeriodicalIF":1.8000,"publicationDate":"2025-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Complexity","FirstCategoryId":"100","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0885064X25000561","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS","Score":null,"Total":0}
引用次数: 0
Abstract
This paper considers deep learning from strongly mixing observations and performs a sparse-penalized regularization for deep neural networks (DNN) predictors. In a general framework that includes regression and classification, oracle inequalities for the expected excess risk are established, and upper bounds on the class of Hölder smooth functions and composition structured Hölder functions are provided. For nonparametric autoregression with the Gaussian and Laplace errors, and the Huber loss function, it is shown that the sparse-penalized DNN estimator proposed is optimal (up to a logarithmic factor) in the minimax sense. Based on the lower bound established in Alquier and Kengne (2024), we show that the proposed DNN estimator for the classification task with the logistic loss on strongly mixing observations achieves (up to a logarithmic factor), the minimax optimal convergence rate.
期刊介绍:
The multidisciplinary Journal of Complexity publishes original research papers that contain substantial mathematical results on complexity as broadly conceived. Outstanding review papers will also be published. In the area of computational complexity, the focus is on complexity over the reals, with the emphasis on lower bounds and optimal algorithms. The Journal of Complexity also publishes articles that provide major new algorithms or make important progress on upper bounds. Other models of computation, such as the Turing machine model, are also of interest. Computational complexity results in a wide variety of areas are solicited.
Areas Include:
• Approximation theory
• Biomedical computing
• Compressed computing and sensing
• Computational finance
• Computational number theory
• Computational stochastics
• Control theory
• Cryptography
• Design of experiments
• Differential equations
• Discrete problems
• Distributed and parallel computation
• High and infinite-dimensional problems
• Information-based complexity
• Inverse and ill-posed problems
• Machine learning
• Markov chain Monte Carlo
• Monte Carlo and quasi-Monte Carlo
• Multivariate integration and approximation
• Noisy data
• Nonlinear and algebraic equations
• Numerical analysis
• Operator equations
• Optimization
• Quantum computing
• Scientific computation
• Tractability of multivariate problems
• Vision and image understanding.