{"title":"Robust Neural Network Training Using Inverted Probability Distribution","authors":"Teerapaun Tanprasert, T. Tanprasert","doi":"10.1145/3426826.3426827","DOIUrl":null,"url":null,"abstract":"This paper presents strategies to tweak the probability distribution of the data set to bias the training process of a neural network for a better learning outcome. For a real-world problem, provided that the probability distribution of the population can be assumed, the training set can be sampled from the population in such a way that its probability distribution satisfies certain targeted characteristics. For example, if the boundary between classes is critical to the training outcome, a larger proportion of training data may be drawn from the area around the boundaries. On the other hand, if the learning outcome is aimed at resembling a common concept encoded in the training set, learning from the data near the norm may be more effective. In order to explore the effectiveness of the various strategies, the concept was applied to two problems: 3-spiral and wine quality. Experimental results suggest that, whether the problem requires an emphasis on classifying boundary or recognizing the central pattern, our novel sampling strategy – inverted probability distribution – performs exceptionally well.","PeriodicalId":202857,"journal":{"name":"Proceedings of the 2020 3rd International Conference on Machine Learning and Machine Intelligence","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2020 3rd International Conference on Machine Learning and Machine Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3426826.3426827","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper presents strategies to tweak the probability distribution of the data set to bias the training process of a neural network for a better learning outcome. For a real-world problem, provided that the probability distribution of the population can be assumed, the training set can be sampled from the population in such a way that its probability distribution satisfies certain targeted characteristics. For example, if the boundary between classes is critical to the training outcome, a larger proportion of training data may be drawn from the area around the boundaries. On the other hand, if the learning outcome is aimed at resembling a common concept encoded in the training set, learning from the data near the norm may be more effective. In order to explore the effectiveness of the various strategies, the concept was applied to two problems: 3-spiral and wine quality. Experimental results suggest that, whether the problem requires an emphasis on classifying boundary or recognizing the central pattern, our novel sampling strategy – inverted probability distribution – performs exceptionally well.