{"title":"判别分析,深度神经网络","authors":"Li Li, M. Doroslovački, M. Loew","doi":"10.1109/CISS.2019.8692803","DOIUrl":null,"url":null,"abstract":"One consensus in the machine learning community is that obtaining good representations of the data is crucial for the classification tasks. But establishing a clear objective for representation learning is an open question and difficult. In this paper, we propose the Discriminant Analysis Loss Function (DALF) for the representation learning in Deep Neural Networks (DNNs). The gradients of DALF explicitly minimize the within-class variances (scatter) and maximize the between-class variances. We use DALF to drive the training of DNNs and call them Discriminant Analysis Deep Neural Networks (DisAnDNNs). Compared to other Linear Discriminant Analysis (LDA)-based cost functions, the computational cost of DALF is drastically reduced by avoiding eigen-decomposition and matrix inversion. We used simple datasets to illustrate the geometric meaning of DALF and compared it with LDA, then experimented with DALF-driven Residual Learning Nets (ResNets) on the pediatric pneumonia (chest X-ray image) dataset. The experimental results show that the DisAnDNNs achieve state-of the-art accuracy in the binary classification task. Particularly, in the pediatric pneumonia dataset, we achieved the accuracy of 96.63%, with a sensitivity of 99.23% and a specificity of 92.30%, all of which are better than the results in the literature that published the dataset.","PeriodicalId":123696,"journal":{"name":"2019 53rd Annual Conference on Information Sciences and Systems (CISS)","volume":"98 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"20","resultStr":"{\"title\":\"Discriminant Analysis Deep Neural Networks\",\"authors\":\"Li Li, M. Doroslovački, M. Loew\",\"doi\":\"10.1109/CISS.2019.8692803\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"One consensus in the machine learning community is that obtaining good representations of the data is crucial for the classification tasks. But establishing a clear objective for representation learning is an open question and difficult. In this paper, we propose the Discriminant Analysis Loss Function (DALF) for the representation learning in Deep Neural Networks (DNNs). The gradients of DALF explicitly minimize the within-class variances (scatter) and maximize the between-class variances. We use DALF to drive the training of DNNs and call them Discriminant Analysis Deep Neural Networks (DisAnDNNs). Compared to other Linear Discriminant Analysis (LDA)-based cost functions, the computational cost of DALF is drastically reduced by avoiding eigen-decomposition and matrix inversion. We used simple datasets to illustrate the geometric meaning of DALF and compared it with LDA, then experimented with DALF-driven Residual Learning Nets (ResNets) on the pediatric pneumonia (chest X-ray image) dataset. The experimental results show that the DisAnDNNs achieve state-of the-art accuracy in the binary classification task. Particularly, in the pediatric pneumonia dataset, we achieved the accuracy of 96.63%, with a sensitivity of 99.23% and a specificity of 92.30%, all of which are better than the results in the literature that published the dataset.\",\"PeriodicalId\":123696,\"journal\":{\"name\":\"2019 53rd Annual Conference on Information Sciences and Systems (CISS)\",\"volume\":\"98 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"20\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 53rd Annual Conference on Information Sciences and Systems (CISS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CISS.2019.8692803\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 53rd Annual Conference on Information Sciences and Systems (CISS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CISS.2019.8692803","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
One consensus in the machine learning community is that obtaining good representations of the data is crucial for the classification tasks. But establishing a clear objective for representation learning is an open question and difficult. In this paper, we propose the Discriminant Analysis Loss Function (DALF) for the representation learning in Deep Neural Networks (DNNs). The gradients of DALF explicitly minimize the within-class variances (scatter) and maximize the between-class variances. We use DALF to drive the training of DNNs and call them Discriminant Analysis Deep Neural Networks (DisAnDNNs). Compared to other Linear Discriminant Analysis (LDA)-based cost functions, the computational cost of DALF is drastically reduced by avoiding eigen-decomposition and matrix inversion. We used simple datasets to illustrate the geometric meaning of DALF and compared it with LDA, then experimented with DALF-driven Residual Learning Nets (ResNets) on the pediatric pneumonia (chest X-ray image) dataset. The experimental results show that the DisAnDNNs achieve state-of the-art accuracy in the binary classification task. Particularly, in the pediatric pneumonia dataset, we achieved the accuracy of 96.63%, with a sensitivity of 99.23% and a specificity of 92.30%, all of which are better than the results in the literature that published the dataset.