Informed machine learning to reconcile interpretability with fidelity in scientific applications

IF 13.9 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Andrea Murari, Riccardo Rossi, Luca Spolladore, Ivan Wyss, Michela Gelfusa
{"title":"Informed machine learning to reconcile interpretability with fidelity in scientific applications","authors":"Andrea Murari,&nbsp;Riccardo Rossi,&nbsp;Luca Spolladore,&nbsp;Ivan Wyss,&nbsp;Michela Gelfusa","doi":"10.1007/s10462-025-11282-y","DOIUrl":null,"url":null,"abstract":"<div><p>Notwithstanding their impressive performances, unfortunately some of the most powerful machine learning (ML) models are obscure and almost impossible to interpret. Consequently, in the last years, there has been a rapid increase in research about eXplainable Artificial Intelligence, whose objective consists of improving their transparency. In scientific applications, explainability assumes a different flavour and cannot be reduced to pure user understanding but there is a premium also on <i>fidelity</i>, on developing models that reflect the actual mechanisms at play in the investigated phenomena. To this end, Genetic Programming supported Symbolic Regression (GPSR), conceived explicitly to manipulate symbols, can present various competitive advantages in finding a good trade-off between interpretability and realism. However, the search spaces are typically too large and the algorithms have to be steered to converge on the desired solutions. The present work describes techniques to constrain GPSR and to combine it with deep learning tools, so that the final models are expressed in terms of interpretable and realistic mathematical equations. The strategies to guide convergence include dimensional analysis, integration of prior information about symmetries and conservation laws, refinements of the fitness function and robust statistics. The performances are improved according to all the main metrics: accuracy, robustness against noise and outliers, capability of handling data sparsity and interpretability. Great attention has been paid to introducing practical solutions, covering most essential aspects of the data analysis process, from the treatment of the uncertainties to the quantification of the equations’ complexity. All the main applications of supervised ML, from regression to classification, are considered (and the extension to unsupervised and reinforcement learning are not expected to pose major difficulties). Theoretical considerations, systematic numerical tests, simulations with multiphysics codes and the results of actual experiments prove the potential of the proposed improvements.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 10","pages":""},"PeriodicalIF":13.9000,"publicationDate":"2025-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11282-y.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Artificial Intelligence Review","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10462-025-11282-y","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Notwithstanding their impressive performances, unfortunately some of the most powerful machine learning (ML) models are obscure and almost impossible to interpret. Consequently, in the last years, there has been a rapid increase in research about eXplainable Artificial Intelligence, whose objective consists of improving their transparency. In scientific applications, explainability assumes a different flavour and cannot be reduced to pure user understanding but there is a premium also on fidelity, on developing models that reflect the actual mechanisms at play in the investigated phenomena. To this end, Genetic Programming supported Symbolic Regression (GPSR), conceived explicitly to manipulate symbols, can present various competitive advantages in finding a good trade-off between interpretability and realism. However, the search spaces are typically too large and the algorithms have to be steered to converge on the desired solutions. The present work describes techniques to constrain GPSR and to combine it with deep learning tools, so that the final models are expressed in terms of interpretable and realistic mathematical equations. The strategies to guide convergence include dimensional analysis, integration of prior information about symmetries and conservation laws, refinements of the fitness function and robust statistics. The performances are improved according to all the main metrics: accuracy, robustness against noise and outliers, capability of handling data sparsity and interpretability. Great attention has been paid to introducing practical solutions, covering most essential aspects of the data analysis process, from the treatment of the uncertainties to the quantification of the equations’ complexity. All the main applications of supervised ML, from regression to classification, are considered (and the extension to unsupervised and reinforcement learning are not expected to pose major difficulties). Theoretical considerations, systematic numerical tests, simulations with multiphysics codes and the results of actual experiments prove the potential of the proposed improvements.

知情机器学习在科学应用中调和可解释性与保真度
尽管它们的表现令人印象深刻,但不幸的是,一些最强大的机器学习(ML)模型是模糊的,几乎不可能解释。因此,在过去几年中,关于可解释人工智能的研究迅速增加,其目标包括提高其透明度。在科学应用中,可解释性具有不同的风格,不能简化为纯粹的用户理解,但保真度也很重要,开发的模型反映了在研究现象中起作用的实际机制。为此,遗传规划支持的符号回归(GPSR)被明确地设想为操纵符号,可以在可解释性和现实性之间找到良好的权衡方面呈现出各种竞争优势。然而,搜索空间通常太大,必须引导算法收敛于期望的解决方案。目前的工作描述了约束GPSR并将其与深度学习工具相结合的技术,以便最终模型用可解释和现实的数学方程表示。引导收敛的策略包括量纲分析、关于对称性和守恒律的先验信息的整合、适应度函数的改进和鲁棒统计。根据所有主要指标:准确性,对噪声和异常值的鲁棒性,处理数据稀疏性和可解释性的能力,性能得到了改进。我们非常注意引入实用的解决方案,涵盖数据分析过程的最基本方面,从不确定性的处理到方程复杂性的量化。考虑了监督式机器学习的所有主要应用,从回归到分类(并且扩展到无监督和强化学习预计不会造成重大困难)。理论分析、系统数值测试、多物理场代码模拟和实际实验结果证明了所提改进的潜力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Artificial Intelligence Review
Artificial Intelligence Review 工程技术-计算机:人工智能
CiteScore
22.00
自引率
3.30%
发文量
194
审稿时长
5.3 months
期刊介绍: Artificial Intelligence Review, a fully open access journal, publishes cutting-edge research in artificial intelligence and cognitive science. It features critical evaluations of applications, techniques, and algorithms, providing a platform for both researchers and application developers. The journal includes refereed survey and tutorial articles, along with reviews and commentary on significant developments in the field.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信