{"title":"Quasi-Equivalence of Width and Depth of Neural Networks","authors":"Fenglei Fan, Rongjie Lai, Ge Wang","doi":"10.21203/rs.3.rs-92324/v1","DOIUrl":null,"url":null,"abstract":"\n While classic studies proved that wide networks allow universal approximation, recent research and successes\nof deep learning demonstrate the power of the network depth. Based on a symmetric consideration,\nwe investigate if the design of artificial neural networks should have a directional preference, and what the\nmechanism of interaction is between the width and depth of a network. We address this fundamental question\nby establishing a quasi-equivalence between the width and depth of ReLU networks. Specifically, we formulate a\ntransformation from an arbitrary ReLU network to a wide network and a deep network for either regression\nor classification so that an essentially same capability of the original network can be implemented. That is, a\ndeep regression/classification ReLU network has a wide equivalent, and vice versa, subject to an arbitrarily small\nerror. Interestingly, the quasi-equivalence between wide and deep classification ReLU networks is a data-driven\nversion of the DeMorgan law.","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"137 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2020-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv: Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.21203/rs.3.rs-92324/v1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 11
Abstract
While classic studies proved that wide networks allow universal approximation, recent research and successes
of deep learning demonstrate the power of the network depth. Based on a symmetric consideration,
we investigate if the design of artificial neural networks should have a directional preference, and what the
mechanism of interaction is between the width and depth of a network. We address this fundamental question
by establishing a quasi-equivalence between the width and depth of ReLU networks. Specifically, we formulate a
transformation from an arbitrary ReLU network to a wide network and a deep network for either regression
or classification so that an essentially same capability of the original network can be implemented. That is, a
deep regression/classification ReLU network has a wide equivalent, and vice versa, subject to an arbitrarily small
error. Interestingly, the quasi-equivalence between wide and deep classification ReLU networks is a data-driven
version of the DeMorgan law.