{"title":"Internal node bagging: a layer-wise ensemble training method","authors":"Jinhong Li, Shun Yi","doi":"10.1145/3357254.3357268","DOIUrl":null,"url":null,"abstract":"When training neural networks, regularization methods are needed to avoid model overfitting. Dropout is a widely used regularization method, but its working principle is inconclusive and it does not work well for small models. This paper introduced a novel view to understand how dropout works as a layer-wise ensemble training method, that each feature in hidden layers is learned by multiple nodes, and next layer integrates the outputs of these nodes. Basing on the novel understanding of dropout, we proposed a new neural network training algorithm named internal node bagging, which explicitly forces a group of nodes to learn the same feature during training phase and combines these nodes into one node during testing phase. This means that more parameters can be used during training phase to improve the fitting ability of models while keeping model remains small during testing phase. After experimenting on three datasets, it is found that this algorithm can significantly improve the test performance of small models.","PeriodicalId":361892,"journal":{"name":"International Conference on Artificial Intelligence and Pattern Recognition","volume":"17 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Artificial Intelligence and Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3357254.3357268","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
When training neural networks, regularization methods are needed to avoid model overfitting. Dropout is a widely used regularization method, but its working principle is inconclusive and it does not work well for small models. This paper introduced a novel view to understand how dropout works as a layer-wise ensemble training method, that each feature in hidden layers is learned by multiple nodes, and next layer integrates the outputs of these nodes. Basing on the novel understanding of dropout, we proposed a new neural network training algorithm named internal node bagging, which explicitly forces a group of nodes to learn the same feature during training phase and combines these nodes into one node during testing phase. This means that more parameters can be used during training phase to improve the fitting ability of models while keeping model remains small during testing phase. After experimenting on three datasets, it is found that this algorithm can significantly improve the test performance of small models.