{"title":"An Empirical Study of Linear Dimensionality Reduction for Judicial Predictive Models","authors":"Zhenyu Liu, Huanhuan Chen","doi":"10.1109/ICIST.2018.8426121","DOIUrl":null,"url":null,"abstract":"Judicial cases can be modeled with thetextual frequency vectors under the Bag-of-Words assumption to predict the decision outcome. However, such models are often with much more numbers of features than training samples, which usually leads to the over fitting problem. In this paper, we conduct an empirical investigation on linear dimensionality reduction of the high-dimensional judicial predictive models via the wide spread principal component analysis approach. The experimental results show that these high-dimensional models do not suffer from the overfitting problem, but the under fitting problem. Moreover, the higher-order dependency in the textual frequency data cannot be decorrelated by the linear dimensionality reduction approach, which restrains the performance of judicial classification models subject to the unchanged level of signal-noise ratio in the derived low-dimensional features.","PeriodicalId":331555,"journal":{"name":"2018 Eighth International Conference on Information Science and Technology (ICIST)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 Eighth International Conference on Information Science and Technology (ICIST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICIST.2018.8426121","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Judicial cases can be modeled with thetextual frequency vectors under the Bag-of-Words assumption to predict the decision outcome. However, such models are often with much more numbers of features than training samples, which usually leads to the over fitting problem. In this paper, we conduct an empirical investigation on linear dimensionality reduction of the high-dimensional judicial predictive models via the wide spread principal component analysis approach. The experimental results show that these high-dimensional models do not suffer from the overfitting problem, but the under fitting problem. Moreover, the higher-order dependency in the textual frequency data cannot be decorrelated by the linear dimensionality reduction approach, which restrains the performance of judicial classification models subject to the unchanged level of signal-noise ratio in the derived low-dimensional features.