{"title":"神经网络宽度和深度的拟等价","authors":"Fenglei Fan, Rongjie Lai, Ge Wang","doi":"10.21203/rs.3.rs-92324/v1","DOIUrl":null,"url":null,"abstract":"\n While classic studies proved that wide networks allow universal approximation, recent research and successes\nof deep learning demonstrate the power of the network depth. Based on a symmetric consideration,\nwe investigate if the design of artificial neural networks should have a directional preference, and what the\nmechanism of interaction is between the width and depth of a network. We address this fundamental question\nby establishing a quasi-equivalence between the width and depth of ReLU networks. Specifically, we formulate a\ntransformation from an arbitrary ReLU network to a wide network and a deep network for either regression\nor classification so that an essentially same capability of the original network can be implemented. That is, a\ndeep regression/classification ReLU network has a wide equivalent, and vice versa, subject to an arbitrarily small\nerror. Interestingly, the quasi-equivalence between wide and deep classification ReLU networks is a data-driven\nversion of the DeMorgan law.","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"137 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2020-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":"{\"title\":\"Quasi-Equivalence of Width and Depth of Neural Networks\",\"authors\":\"Fenglei Fan, Rongjie Lai, Ge Wang\",\"doi\":\"10.21203/rs.3.rs-92324/v1\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"\\n While classic studies proved that wide networks allow universal approximation, recent research and successes\\nof deep learning demonstrate the power of the network depth. Based on a symmetric consideration,\\nwe investigate if the design of artificial neural networks should have a directional preference, and what the\\nmechanism of interaction is between the width and depth of a network. We address this fundamental question\\nby establishing a quasi-equivalence between the width and depth of ReLU networks. Specifically, we formulate a\\ntransformation from an arbitrary ReLU network to a wide network and a deep network for either regression\\nor classification so that an essentially same capability of the original network can be implemented. That is, a\\ndeep regression/classification ReLU network has a wide equivalent, and vice versa, subject to an arbitrarily small\\nerror. Interestingly, the quasi-equivalence between wide and deep classification ReLU networks is a data-driven\\nversion of the DeMorgan law.\",\"PeriodicalId\":8468,\"journal\":{\"name\":\"arXiv: Learning\",\"volume\":\"137 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-02-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"11\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv: Learning\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.21203/rs.3.rs-92324/v1\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv: Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.21203/rs.3.rs-92324/v1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Quasi-Equivalence of Width and Depth of Neural Networks
While classic studies proved that wide networks allow universal approximation, recent research and successes
of deep learning demonstrate the power of the network depth. Based on a symmetric consideration,
we investigate if the design of artificial neural networks should have a directional preference, and what the
mechanism of interaction is between the width and depth of a network. We address this fundamental question
by establishing a quasi-equivalence between the width and depth of ReLU networks. Specifically, we formulate a
transformation from an arbitrary ReLU network to a wide network and a deep network for either regression
or classification so that an essentially same capability of the original network can be implemented. That is, a
deep regression/classification ReLU network has a wide equivalent, and vice versa, subject to an arbitrarily small
error. Interestingly, the quasi-equivalence between wide and deep classification ReLU networks is a data-driven
version of the DeMorgan law.