{"title":"理论二:深度学习和优化","authors":"T. Poggio, Q. Liao","doi":"10.24425/BPAS.2018.125925","DOIUrl":null,"url":null,"abstract":"Bull. Pol. Ac.: Tech. 66(6) 2018 Abstract. The landscape of the empirical risk of overparametrized deep convolutional neural networks (DCNNs) is characterized with a mix of theory and experiments. In part A we show the existence of a large number of global minimizers with zero empirical error (modulo inconsistent equations). The argument which relies on the use of Bezout theorem is rigorous when the RELUs are replaced by a polynomial nonlinearity. We show with simulations that the corresponding polynomial network is indistinguishable from the RELU network. According to Bezout theorem, the global minimizers are degenerate unlike the local minima which in general should be non-degenerate. Further we experimentally analyzed and visualized the landscape of empirical risk of DCNNs on CIFAR-10 dataset. Based on above theoretical and experimental observations, we propose a simple model of the landscape of empirical risk. In part B, we characterize the optimization properties of stochastic gradient descent applied to deep networks. The main claim here consists of theoretical and experimental evidence for the following property of SGD: SGD concentrates in probability – like the classical Langevin equation – on large volume, ”flat” minima, selecting with high probability degenerate minimizers which are typically global minimizers.","PeriodicalId":55299,"journal":{"name":"Bulletin of the Polish Academy of Sciences-Technical Sciences","volume":"9 1","pages":""},"PeriodicalIF":1.2000,"publicationDate":"2023-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":"{\"title\":\"Theory II: Deep learning and optimization\",\"authors\":\"T. Poggio, Q. Liao\",\"doi\":\"10.24425/BPAS.2018.125925\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Bull. Pol. Ac.: Tech. 66(6) 2018 Abstract. The landscape of the empirical risk of overparametrized deep convolutional neural networks (DCNNs) is characterized with a mix of theory and experiments. In part A we show the existence of a large number of global minimizers with zero empirical error (modulo inconsistent equations). The argument which relies on the use of Bezout theorem is rigorous when the RELUs are replaced by a polynomial nonlinearity. We show with simulations that the corresponding polynomial network is indistinguishable from the RELU network. According to Bezout theorem, the global minimizers are degenerate unlike the local minima which in general should be non-degenerate. Further we experimentally analyzed and visualized the landscape of empirical risk of DCNNs on CIFAR-10 dataset. Based on above theoretical and experimental observations, we propose a simple model of the landscape of empirical risk. In part B, we characterize the optimization properties of stochastic gradient descent applied to deep networks. The main claim here consists of theoretical and experimental evidence for the following property of SGD: SGD concentrates in probability – like the classical Langevin equation – on large volume, ”flat” minima, selecting with high probability degenerate minimizers which are typically global minimizers.\",\"PeriodicalId\":55299,\"journal\":{\"name\":\"Bulletin of the Polish Academy of Sciences-Technical Sciences\",\"volume\":\"9 1\",\"pages\":\"\"},\"PeriodicalIF\":1.2000,\"publicationDate\":\"2023-11-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"8\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Bulletin of the Polish Academy of Sciences-Technical Sciences\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.24425/BPAS.2018.125925\",\"RegionNum\":4,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"ENGINEERING, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Bulletin of the Polish Academy of Sciences-Technical Sciences","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.24425/BPAS.2018.125925","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
Bull. Pol. Ac.: Tech. 66(6) 2018 Abstract. The landscape of the empirical risk of overparametrized deep convolutional neural networks (DCNNs) is characterized with a mix of theory and experiments. In part A we show the existence of a large number of global minimizers with zero empirical error (modulo inconsistent equations). The argument which relies on the use of Bezout theorem is rigorous when the RELUs are replaced by a polynomial nonlinearity. We show with simulations that the corresponding polynomial network is indistinguishable from the RELU network. According to Bezout theorem, the global minimizers are degenerate unlike the local minima which in general should be non-degenerate. Further we experimentally analyzed and visualized the landscape of empirical risk of DCNNs on CIFAR-10 dataset. Based on above theoretical and experimental observations, we propose a simple model of the landscape of empirical risk. In part B, we characterize the optimization properties of stochastic gradient descent applied to deep networks. The main claim here consists of theoretical and experimental evidence for the following property of SGD: SGD concentrates in probability – like the classical Langevin equation – on large volume, ”flat” minima, selecting with high probability degenerate minimizers which are typically global minimizers.
期刊介绍:
The Bulletin of the Polish Academy of Sciences: Technical Sciences is published bimonthly by the Division IV Engineering Sciences of the Polish Academy of Sciences, since the beginning of the existence of the PAS in 1952. The journal is peer‐reviewed and is published both in printed and electronic form. It is established for the publication of original high quality papers from multidisciplinary Engineering sciences with the following topics preferred:
Artificial and Computational Intelligence,
Biomedical Engineering and Biotechnology,
Civil Engineering,
Control, Informatics and Robotics,
Electronics, Telecommunication and Optoelectronics,
Mechanical and Aeronautical Engineering, Thermodynamics,
Material Science and Nanotechnology,
Power Systems and Power Electronics.