{"title":"Neural Network Image Classifiers Informed by Factor Analyzers","authors":"A. M. Dostovalova, A. K. Gorshenin","doi":"10.1134/S106456242460204X","DOIUrl":null,"url":null,"abstract":"<p>The paper develops an approach to probability informing deep neural networks, that is, improving their results by using various probability models within architectural elements. We introduce factor analyzers with additive and impulse noise components as such models. The identifiability of the model is proved. The relationship between the parameter estimates by the methods of least squares and maximum likelihood is established, which actually means that the estimates of the parameters of the factor analyzer obtained within the informed block are unbiased and consistent. A mathematical model is used to create a new architectural element that implements the fusion of multiscale image features to improve classification accuracy in the case of a small volume of training data. This problem is typical for various applied tasks, including remote sensing data analysis. Various widely used neural network classifiers (EfficientNet, MobileNet, and Xception), both with and without a new informed block, are tested. It is demonstrated that on the open datasets UC Merced (remote sensing data) and Oxford Flowers (flower images), informed neural networks achieve a significant increase in accuracy for this class of tasks: the largest improvement in Top-1 accuracy was 6.67% (mean accuracy without informing equals 87.3%), while Top-5 accuracy increased by 1.49% (mean base accuracy value is 96.27%).</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S35 - S41"},"PeriodicalIF":0.5000,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S106456242460204X.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Doklady Mathematics","FirstCategoryId":"100","ListUrlMain":"https://link.springer.com/article/10.1134/S106456242460204X","RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"MATHEMATICS","Score":null,"Total":0}
引用次数: 0
Abstract
The paper develops an approach to probability informing deep neural networks, that is, improving their results by using various probability models within architectural elements. We introduce factor analyzers with additive and impulse noise components as such models. The identifiability of the model is proved. The relationship between the parameter estimates by the methods of least squares and maximum likelihood is established, which actually means that the estimates of the parameters of the factor analyzer obtained within the informed block are unbiased and consistent. A mathematical model is used to create a new architectural element that implements the fusion of multiscale image features to improve classification accuracy in the case of a small volume of training data. This problem is typical for various applied tasks, including remote sensing data analysis. Various widely used neural network classifiers (EfficientNet, MobileNet, and Xception), both with and without a new informed block, are tested. It is demonstrated that on the open datasets UC Merced (remote sensing data) and Oxford Flowers (flower images), informed neural networks achieve a significant increase in accuracy for this class of tasks: the largest improvement in Top-1 accuracy was 6.67% (mean accuracy without informing equals 87.3%), while Top-5 accuracy increased by 1.49% (mean base accuracy value is 96.27%).
期刊介绍:
Doklady Mathematics is a journal of the Presidium of the Russian Academy of Sciences. It contains English translations of papers published in Doklady Akademii Nauk (Proceedings of the Russian Academy of Sciences), which was founded in 1933 and is published 36 times a year. Doklady Mathematics includes the materials from the following areas: mathematics, mathematical physics, computer science, control theory, and computers. It publishes brief scientific reports on previously unpublished significant new research in mathematics and its applications. The main contributors to the journal are Members of the RAS, Corresponding Members of the RAS, and scientists from the former Soviet Union and other foreign countries. Among the contributors are the outstanding Russian mathematicians.