{"title":"Interpretable logical-probabilistic approximation of neural networks","authors":"Evgenii Vityaev , Alexey Korolev","doi":"10.1016/j.cogsys.2024.101301","DOIUrl":null,"url":null,"abstract":"<div><div>The paper proposes the approximation of DNNs by replacing each neuron by the corresponding logical-probabilistic neuron. Logical-probabilistic neurons learn their behavior based on the responses of initial neurons on incoming signals and discover all logical-probabilistic causal relationships between the input and output. These logical-probabilistic causal relationships are, in a certain sense, most precise – it was proved in the previous works that they are theoretically (when probability is known) can predict without contradictions. The resulting logical-probabilistic neurons are interconnected by the same connections as the initial neurons after replacing their signals on true/false. The resulting logical-probabilistic neural network produces its own predictions that approximate the predictions of the original DNN. Thus, we obtain an interpretable approximation of DNN, which also allows tracing of DNN by tracing its excitations through the causal relationships. This approximation of DNN is a Distillation method such as Model Translation, which train alternative smaller interpretable models that mimics the total input/output behavior of DNN. It is also locally interpretable and explains every particular prediction. It explains the sequences of logical probabilistic causal relationships that infer that prediction and also show all features that took part in this prediction with the statistical estimation of their significance. Experimental results on approximation accuracy of all intermedia neurons, output neurons and softmax output of DNN are presented, as well as the accuracy of obtained logical-probabilistic neural network. From the practical point of view, interpretable transformation of neural networks is very important for the hybrid artificial intelligent systems, where neural networks are integrated with the symbolic methods of AI. As a practical application we consider smart city.</div></div>","PeriodicalId":55242,"journal":{"name":"Cognitive Systems Research","volume":"88 ","pages":"Article 101301"},"PeriodicalIF":2.1000,"publicationDate":"2024-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Cognitive Systems Research","FirstCategoryId":"102","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1389041724000950","RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The paper proposes the approximation of DNNs by replacing each neuron by the corresponding logical-probabilistic neuron. Logical-probabilistic neurons learn their behavior based on the responses of initial neurons on incoming signals and discover all logical-probabilistic causal relationships between the input and output. These logical-probabilistic causal relationships are, in a certain sense, most precise – it was proved in the previous works that they are theoretically (when probability is known) can predict without contradictions. The resulting logical-probabilistic neurons are interconnected by the same connections as the initial neurons after replacing their signals on true/false. The resulting logical-probabilistic neural network produces its own predictions that approximate the predictions of the original DNN. Thus, we obtain an interpretable approximation of DNN, which also allows tracing of DNN by tracing its excitations through the causal relationships. This approximation of DNN is a Distillation method such as Model Translation, which train alternative smaller interpretable models that mimics the total input/output behavior of DNN. It is also locally interpretable and explains every particular prediction. It explains the sequences of logical probabilistic causal relationships that infer that prediction and also show all features that took part in this prediction with the statistical estimation of their significance. Experimental results on approximation accuracy of all intermedia neurons, output neurons and softmax output of DNN are presented, as well as the accuracy of obtained logical-probabilistic neural network. From the practical point of view, interpretable transformation of neural networks is very important for the hybrid artificial intelligent systems, where neural networks are integrated with the symbolic methods of AI. As a practical application we consider smart city.
期刊介绍:
Cognitive Systems Research is dedicated to the study of human-level cognition. As such, it welcomes papers which advance the understanding, design and applications of cognitive and intelligent systems, both natural and artificial.
The journal brings together a broad community studying cognition in its many facets in vivo and in silico, across the developmental spectrum, focusing on individual capacities or on entire architectures. It aims to foster debate and integrate ideas, concepts, constructs, theories, models and techniques from across different disciplines and different perspectives on human-level cognition. The scope of interest includes the study of cognitive capacities and architectures - both brain-inspired and non-brain-inspired - and the application of cognitive systems to real-world problems as far as it offers insights relevant for the understanding of cognition.
Cognitive Systems Research therefore welcomes mature and cutting-edge research approaching cognition from a systems-oriented perspective, both theoretical and empirically-informed, in the form of original manuscripts, short communications, opinion articles, systematic reviews, and topical survey articles from the fields of Cognitive Science (including Philosophy of Cognitive Science), Artificial Intelligence/Computer Science, Cognitive Robotics, Developmental Science, Psychology, and Neuroscience and Neuromorphic Engineering. Empirical studies will be considered if they are supplemented by theoretical analyses and contributions to theory development and/or computational modelling studies.