{"title":"Towards machine learning as AGM-style belief change","authors":"Theofanis Aravanis","doi":"10.1016/j.ijar.2025.109437","DOIUrl":null,"url":null,"abstract":"<div><div>Artificial Neural Networks (ANNs) are powerful computational models that are able to reproduce complex non-linear processes, and are being widely used in a plethora of contemporary disciplines. In this article, we study the statics and dynamics of a certain class of ANNs, called binary ANNs, from the perspective of belief-change theory. A binary ANN is a feed-forward ANN whose inputs and outputs take binary values, and as such, it is suitable for a wide range of practical applications. For this type of ANNs, we point out that their knowledge (expressed via their input-output relationship) can symbolically be represented in terms of a propositional logic language. Furthermore, in the realm of belief change, we identify the process of changing (revising/contracting) an initial belief set to a modified belief set, as a process of a gradual transition of intermediate belief sets — such a gradualist approach to belief change is more congruent with the behaviors of real-world agents. Along these lines, we provide natural metrics for measuring the distance between these intermediate belief sets, effectively quantifying the disparity in their encoded knowledge. Thereafter, we demonstrate that, similar to belief change, the training process of binary ANNs, through backpropagation, can be emulated via a sequence of successive transitions of belief sets, the distance between which is intuitively related through one of the aforementioned metrics. We also prove that the alluded successive transitions of belief sets can be modeled by means of rational revision and contraction operators, defined within the fundamental belief-change framework of Alchourrón, Gärdenfors and Makinson (AGM). Thus, the process of machine learning (specifically, training binary ANNs) is framed as an operation of AGM-style belief change, offering a modular and logically structured perspective on neural learning.</div></div>","PeriodicalId":13842,"journal":{"name":"International Journal of Approximate Reasoning","volume":"183 ","pages":"Article 109437"},"PeriodicalIF":3.2000,"publicationDate":"2025-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Approximate Reasoning","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0888613X25000787","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Artificial Neural Networks (ANNs) are powerful computational models that are able to reproduce complex non-linear processes, and are being widely used in a plethora of contemporary disciplines. In this article, we study the statics and dynamics of a certain class of ANNs, called binary ANNs, from the perspective of belief-change theory. A binary ANN is a feed-forward ANN whose inputs and outputs take binary values, and as such, it is suitable for a wide range of practical applications. For this type of ANNs, we point out that their knowledge (expressed via their input-output relationship) can symbolically be represented in terms of a propositional logic language. Furthermore, in the realm of belief change, we identify the process of changing (revising/contracting) an initial belief set to a modified belief set, as a process of a gradual transition of intermediate belief sets — such a gradualist approach to belief change is more congruent with the behaviors of real-world agents. Along these lines, we provide natural metrics for measuring the distance between these intermediate belief sets, effectively quantifying the disparity in their encoded knowledge. Thereafter, we demonstrate that, similar to belief change, the training process of binary ANNs, through backpropagation, can be emulated via a sequence of successive transitions of belief sets, the distance between which is intuitively related through one of the aforementioned metrics. We also prove that the alluded successive transitions of belief sets can be modeled by means of rational revision and contraction operators, defined within the fundamental belief-change framework of Alchourrón, Gärdenfors and Makinson (AGM). Thus, the process of machine learning (specifically, training binary ANNs) is framed as an operation of AGM-style belief change, offering a modular and logically structured perspective on neural learning.
期刊介绍:
The International Journal of Approximate Reasoning is intended to serve as a forum for the treatment of imprecision and uncertainty in Artificial and Computational Intelligence, covering both the foundations of uncertainty theories, and the design of intelligent systems for scientific and engineering applications. It publishes high-quality research papers describing theoretical developments or innovative applications, as well as review articles on topics of general interest.
Relevant topics include, but are not limited to, probabilistic reasoning and Bayesian networks, imprecise probabilities, random sets, belief functions (Dempster-Shafer theory), possibility theory, fuzzy sets, rough sets, decision theory, non-additive measures and integrals, qualitative reasoning about uncertainty, comparative probability orderings, game-theoretic probability, default reasoning, nonstandard logics, argumentation systems, inconsistency tolerant reasoning, elicitation techniques, philosophical foundations and psychological models of uncertain reasoning.
Domains of application for uncertain reasoning systems include risk analysis and assessment, information retrieval and database design, information fusion, machine learning, data and web mining, computer vision, image and signal processing, intelligent data analysis, statistics, multi-agent systems, etc.