{"title":"Deep Neural Networks: Multi-Classification and Universal Approximation","authors":"Martín Hernández, Enrique Zuazua","doi":"arxiv-2409.06555","DOIUrl":null,"url":null,"abstract":"We demonstrate that a ReLU deep neural network with a width of $2$ and a\ndepth of $2N+4M-1$ layers can achieve finite sample memorization for any\ndataset comprising $N$ elements in $\\mathbb{R}^d$, where $d\\ge1,$ and $M$\nclasses, thereby ensuring accurate classification. By modeling the neural network as a time-discrete nonlinear dynamical system,\nwe interpret the memorization property as a problem of simultaneous or ensemble\ncontrollability. This problem is addressed by constructing the network\nparameters inductively and explicitly, bypassing the need for training or\nsolving any optimization problem. Additionally, we establish that such a network can achieve universal\napproximation in $L^p(\\Omega;\\mathbb{R}_+)$, where $\\Omega$ is a bounded subset\nof $\\mathbb{R}^d$ and $p\\in[1,\\infty)$, using a ReLU deep neural network with a\nwidth of $d+1$. We also provide depth estimates for approximating $W^{1,p}$\nfunctions and width estimates for approximating $L^p(\\Omega;\\mathbb{R}^m)$ for\n$m\\geq1$. Our proofs are constructive, offering explicit values for the biases\nand weights involved.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":"3 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.06555","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We demonstrate that a ReLU deep neural network with a width of $2$ and a
depth of $2N+4M-1$ layers can achieve finite sample memorization for any
dataset comprising $N$ elements in $\mathbb{R}^d$, where $d\ge1,$ and $M$
classes, thereby ensuring accurate classification. By modeling the neural network as a time-discrete nonlinear dynamical system,
we interpret the memorization property as a problem of simultaneous or ensemble
controllability. This problem is addressed by constructing the network
parameters inductively and explicitly, bypassing the need for training or
solving any optimization problem. Additionally, we establish that such a network can achieve universal
approximation in $L^p(\Omega;\mathbb{R}_+)$, where $\Omega$ is a bounded subset
of $\mathbb{R}^d$ and $p\in[1,\infty)$, using a ReLU deep neural network with a
width of $d+1$. We also provide depth estimates for approximating $W^{1,p}$
functions and width estimates for approximating $L^p(\Omega;\mathbb{R}^m)$ for
$m\geq1$. Our proofs are constructive, offering explicit values for the biases
and weights involved.