{"title":"ReLU浅层神经网络的近似速率","authors":"Tong Mao , Ding-Xuan Zhou","doi":"10.1016/j.jco.2023.101784","DOIUrl":null,"url":null,"abstract":"<div><p>Neural networks activated by the rectified linear unit (ReLU) play a central role in the recent development of deep learning. The topic of approximating functions from Hölder spaces by these networks is crucial for understanding the efficiency of the induced learning algorithms. Although the topic has been well investigated in the setting of deep neural networks with many layers of hidden neurons, it is still open for shallow networks having only one hidden layer. In this paper, we provide rates of uniform approximation by these networks. We show that ReLU shallow neural networks with <em>m</em> hidden neurons can uniformly approximate functions from the Hölder space <span><math><msubsup><mrow><mi>W</mi></mrow><mrow><mo>∞</mo></mrow><mrow><mi>r</mi></mrow></msubsup><mo>(</mo><msup><mrow><mo>[</mo><mo>−</mo><mn>1</mn><mo>,</mo><mn>1</mn><mo>]</mo></mrow><mrow><mi>d</mi></mrow></msup><mo>)</mo></math></span> with rates <span><math><mi>O</mi><mo>(</mo><msup><mrow><mo>(</mo><mi>log</mi><mo></mo><mi>m</mi><mo>)</mo></mrow><mrow><mfrac><mrow><mn>1</mn></mrow><mrow><mn>2</mn></mrow></mfrac><mo>+</mo><mi>d</mi></mrow></msup><msup><mrow><mi>m</mi></mrow><mrow><mo>−</mo><mfrac><mrow><mi>r</mi></mrow><mrow><mi>d</mi></mrow></mfrac><mfrac><mrow><mi>d</mi><mo>+</mo><mn>2</mn></mrow><mrow><mi>d</mi><mo>+</mo><mn>4</mn></mrow></mfrac></mrow></msup><mo>)</mo></math></span> when <span><math><mi>r</mi><mo><</mo><mi>d</mi><mo>/</mo><mn>2</mn><mo>+</mo><mn>2</mn></math></span>. Such rates are very close to the optimal one <span><math><mi>O</mi><mo>(</mo><msup><mrow><mi>m</mi></mrow><mrow><mo>−</mo><mfrac><mrow><mi>r</mi></mrow><mrow><mi>d</mi></mrow></mfrac></mrow></msup><mo>)</mo></math></span> in the sense that <span><math><mfrac><mrow><mi>d</mi><mo>+</mo><mn>2</mn></mrow><mrow><mi>d</mi><mo>+</mo><mn>4</mn></mrow></mfrac></math></span> is close to 1, when the dimension <em>d</em> is large.</p></div>","PeriodicalId":50227,"journal":{"name":"Journal of Complexity","volume":"79 ","pages":"Article 101784"},"PeriodicalIF":1.8000,"publicationDate":"2023-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Rates of approximation by ReLU shallow neural networks\",\"authors\":\"Tong Mao , Ding-Xuan Zhou\",\"doi\":\"10.1016/j.jco.2023.101784\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Neural networks activated by the rectified linear unit (ReLU) play a central role in the recent development of deep learning. The topic of approximating functions from Hölder spaces by these networks is crucial for understanding the efficiency of the induced learning algorithms. Although the topic has been well investigated in the setting of deep neural networks with many layers of hidden neurons, it is still open for shallow networks having only one hidden layer. In this paper, we provide rates of uniform approximation by these networks. We show that ReLU shallow neural networks with <em>m</em> hidden neurons can uniformly approximate functions from the Hölder space <span><math><msubsup><mrow><mi>W</mi></mrow><mrow><mo>∞</mo></mrow><mrow><mi>r</mi></mrow></msubsup><mo>(</mo><msup><mrow><mo>[</mo><mo>−</mo><mn>1</mn><mo>,</mo><mn>1</mn><mo>]</mo></mrow><mrow><mi>d</mi></mrow></msup><mo>)</mo></math></span> with rates <span><math><mi>O</mi><mo>(</mo><msup><mrow><mo>(</mo><mi>log</mi><mo></mo><mi>m</mi><mo>)</mo></mrow><mrow><mfrac><mrow><mn>1</mn></mrow><mrow><mn>2</mn></mrow></mfrac><mo>+</mo><mi>d</mi></mrow></msup><msup><mrow><mi>m</mi></mrow><mrow><mo>−</mo><mfrac><mrow><mi>r</mi></mrow><mrow><mi>d</mi></mrow></mfrac><mfrac><mrow><mi>d</mi><mo>+</mo><mn>2</mn></mrow><mrow><mi>d</mi><mo>+</mo><mn>4</mn></mrow></mfrac></mrow></msup><mo>)</mo></math></span> when <span><math><mi>r</mi><mo><</mo><mi>d</mi><mo>/</mo><mn>2</mn><mo>+</mo><mn>2</mn></math></span>. Such rates are very close to the optimal one <span><math><mi>O</mi><mo>(</mo><msup><mrow><mi>m</mi></mrow><mrow><mo>−</mo><mfrac><mrow><mi>r</mi></mrow><mrow><mi>d</mi></mrow></mfrac></mrow></msup><mo>)</mo></math></span> in the sense that <span><math><mfrac><mrow><mi>d</mi><mo>+</mo><mn>2</mn></mrow><mrow><mi>d</mi><mo>+</mo><mn>4</mn></mrow></mfrac></math></span> is close to 1, when the dimension <em>d</em> is large.</p></div>\",\"PeriodicalId\":50227,\"journal\":{\"name\":\"Journal of Complexity\",\"volume\":\"79 \",\"pages\":\"Article 101784\"},\"PeriodicalIF\":1.8000,\"publicationDate\":\"2023-07-31\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Complexity\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0885064X23000535\",\"RegionNum\":2,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Complexity","FirstCategoryId":"100","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0885064X23000535","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS","Score":null,"Total":0}
Rates of approximation by ReLU shallow neural networks
Neural networks activated by the rectified linear unit (ReLU) play a central role in the recent development of deep learning. The topic of approximating functions from Hölder spaces by these networks is crucial for understanding the efficiency of the induced learning algorithms. Although the topic has been well investigated in the setting of deep neural networks with many layers of hidden neurons, it is still open for shallow networks having only one hidden layer. In this paper, we provide rates of uniform approximation by these networks. We show that ReLU shallow neural networks with m hidden neurons can uniformly approximate functions from the Hölder space with rates when . Such rates are very close to the optimal one in the sense that is close to 1, when the dimension d is large.
期刊介绍:
The multidisciplinary Journal of Complexity publishes original research papers that contain substantial mathematical results on complexity as broadly conceived. Outstanding review papers will also be published. In the area of computational complexity, the focus is on complexity over the reals, with the emphasis on lower bounds and optimal algorithms. The Journal of Complexity also publishes articles that provide major new algorithms or make important progress on upper bounds. Other models of computation, such as the Turing machine model, are also of interest. Computational complexity results in a wide variety of areas are solicited.
Areas Include:
• Approximation theory
• Biomedical computing
• Compressed computing and sensing
• Computational finance
• Computational number theory
• Computational stochastics
• Control theory
• Cryptography
• Design of experiments
• Differential equations
• Discrete problems
• Distributed and parallel computation
• High and infinite-dimensional problems
• Information-based complexity
• Inverse and ill-posed problems
• Machine learning
• Markov chain Monte Carlo
• Monte Carlo and quasi-Monte Carlo
• Multivariate integration and approximation
• Noisy data
• Nonlinear and algebraic equations
• Numerical analysis
• Operator equations
• Optimization
• Quantum computing
• Scientific computation
• Tractability of multivariate problems
• Vision and image understanding.