{"title":"An Empirical Study on Unsupervised Pre-training Approaches in Regression Problems","authors":"P. Saikia, R. Baruah","doi":"10.1109/SSCI.2018.8628674","DOIUrl":null,"url":null,"abstract":"Unsupervised pre-training allows for efficient training of deep architectures. It provides a good set of initialised weights to the deep architecture that can provide better generalisation of the data. In this paper, we aim to empirically analyse the effect of different unsupervised pre-training approaches for the task of regression on different datasets. We have considered two most common pre-training methods namely deep belief network and stacked autoencoder, and compared the results with the standard training algorithm without pretraining. The models with pretraining performed better than the model without pretraining in terms of error, convergence and the prediction of pattern. The results of the experiments also show the importance of hyperparameters tuning, specially learning rate, in providing a better prediction result. This study once again confirmed the effectiveness and potential of pretraining approach in nonlinear regression problem.","PeriodicalId":235735,"journal":{"name":"2018 IEEE Symposium Series on Computational Intelligence (SSCI)","volume":"144 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE Symposium Series on Computational Intelligence (SSCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SSCI.2018.8628674","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Unsupervised pre-training allows for efficient training of deep architectures. It provides a good set of initialised weights to the deep architecture that can provide better generalisation of the data. In this paper, we aim to empirically analyse the effect of different unsupervised pre-training approaches for the task of regression on different datasets. We have considered two most common pre-training methods namely deep belief network and stacked autoencoder, and compared the results with the standard training algorithm without pretraining. The models with pretraining performed better than the model without pretraining in terms of error, convergence and the prediction of pattern. The results of the experiments also show the importance of hyperparameters tuning, specially learning rate, in providing a better prediction result. This study once again confirmed the effectiveness and potential of pretraining approach in nonlinear regression problem.