复杂网络动态神经网络模型的泛化

IF 5.4 1区 物理与天体物理 Q1 PHYSICS, MULTIDISCIPLINARY
Vaiva Vasiliauskaite, Nino Antulov-Fantulin
{"title":"复杂网络动态神经网络模型的泛化","authors":"Vaiva Vasiliauskaite, Nino Antulov-Fantulin","doi":"10.1038/s42005-024-01837-w","DOIUrl":null,"url":null,"abstract":"Differential equations are a ubiquitous tool to study dynamics, ranging from physical systems to complex systems, where a large number of agents interact through a graph. Data-driven approximations of differential equations present a promising alternative to traditional methods for uncovering a model of dynamical systems, especially in complex systems that lack explicit first principles. A recently employed machine learning tool for studying dynamics is neural networks, which can be used for solution finding or discovery of differential equations. However, deploying deep learning models in unfamiliar settings-such as predicting dynamics in unobserved state space regions or on novel graphs-can lead to spurious results. Focusing on complex systems whose dynamics are described with a system of first-order differential equations coupled through a graph, we study generalization of neural network predictions in settings where statistical properties of test data and training data are different. We find that neural networks can accurately predict dynamics beyond the immediate training setting within the domain of the training data. To identify when a model is unable to generalize to novel settings, we propose a statistical significance test. Deep learning is a promising alternative to traditional methods for discovering governing equations, such as variational and perturbation methods, or data-driven approaches like symbolic regression. This paper explores the generalization of neural approximations of dynamics on complex networks to novel, unobserved settings and proposes a statistical testing framework to quantify confidence in the inferred predictions.","PeriodicalId":10540,"journal":{"name":"Communications Physics","volume":" ","pages":"1-10"},"PeriodicalIF":5.4000,"publicationDate":"2024-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.com/articles/s42005-024-01837-w.pdf","citationCount":"0","resultStr":"{\"title\":\"Generalization of neural network models for complex network dynamics\",\"authors\":\"Vaiva Vasiliauskaite, Nino Antulov-Fantulin\",\"doi\":\"10.1038/s42005-024-01837-w\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Differential equations are a ubiquitous tool to study dynamics, ranging from physical systems to complex systems, where a large number of agents interact through a graph. Data-driven approximations of differential equations present a promising alternative to traditional methods for uncovering a model of dynamical systems, especially in complex systems that lack explicit first principles. A recently employed machine learning tool for studying dynamics is neural networks, which can be used for solution finding or discovery of differential equations. However, deploying deep learning models in unfamiliar settings-such as predicting dynamics in unobserved state space regions or on novel graphs-can lead to spurious results. Focusing on complex systems whose dynamics are described with a system of first-order differential equations coupled through a graph, we study generalization of neural network predictions in settings where statistical properties of test data and training data are different. We find that neural networks can accurately predict dynamics beyond the immediate training setting within the domain of the training data. To identify when a model is unable to generalize to novel settings, we propose a statistical significance test. Deep learning is a promising alternative to traditional methods for discovering governing equations, such as variational and perturbation methods, or data-driven approaches like symbolic regression. This paper explores the generalization of neural approximations of dynamics on complex networks to novel, unobserved settings and proposes a statistical testing framework to quantify confidence in the inferred predictions.\",\"PeriodicalId\":10540,\"journal\":{\"name\":\"Communications Physics\",\"volume\":\" \",\"pages\":\"1-10\"},\"PeriodicalIF\":5.4000,\"publicationDate\":\"2024-10-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.nature.com/articles/s42005-024-01837-w.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Communications Physics\",\"FirstCategoryId\":\"101\",\"ListUrlMain\":\"https://www.nature.com/articles/s42005-024-01837-w\",\"RegionNum\":1,\"RegionCategory\":\"物理与天体物理\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PHYSICS, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Communications Physics","FirstCategoryId":"101","ListUrlMain":"https://www.nature.com/articles/s42005-024-01837-w","RegionNum":1,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PHYSICS, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

摘要

微分方程是研究从物理系统到复杂系统等各种动力学的普遍工具。数据驱动的微分方程近似是一种很有前途的替代传统方法,可用于揭示动态系统模型,尤其是在缺乏明确第一原理的复杂系统中。最近用于研究动力学的一种机器学习工具是神经网络,它可用于微分方程的求解或发现。然而,在不熟悉的环境中部署深度学习模型,如预测未观察状态空间区域或新图形上的动态,可能会导致虚假结果。我们以复杂系统为重点,研究了在测试数据和训练数据的统计属性不同的情况下,神经网络预测的泛化。我们发现,在训练数据的范围内,神经网络可以准确预测直接训练环境之外的动态。为了确定模型何时无法泛化到新环境中,我们提出了一种统计显著性检验方法。深度学习是发现控制方程的传统方法(如变分法和扰动法)或数据驱动方法(如符号回归法)的一种有前途的替代方法。本文探讨了复杂网络动力学的神经近似如何泛化到新的、未观察到的环境中,并提出了一个统计检验框架来量化推断预测的置信度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

Generalization of neural network models for complex network dynamics

Generalization of neural network models for complex network dynamics
Differential equations are a ubiquitous tool to study dynamics, ranging from physical systems to complex systems, where a large number of agents interact through a graph. Data-driven approximations of differential equations present a promising alternative to traditional methods for uncovering a model of dynamical systems, especially in complex systems that lack explicit first principles. A recently employed machine learning tool for studying dynamics is neural networks, which can be used for solution finding or discovery of differential equations. However, deploying deep learning models in unfamiliar settings-such as predicting dynamics in unobserved state space regions or on novel graphs-can lead to spurious results. Focusing on complex systems whose dynamics are described with a system of first-order differential equations coupled through a graph, we study generalization of neural network predictions in settings where statistical properties of test data and training data are different. We find that neural networks can accurately predict dynamics beyond the immediate training setting within the domain of the training data. To identify when a model is unable to generalize to novel settings, we propose a statistical significance test. Deep learning is a promising alternative to traditional methods for discovering governing equations, such as variational and perturbation methods, or data-driven approaches like symbolic regression. This paper explores the generalization of neural approximations of dynamics on complex networks to novel, unobserved settings and proposes a statistical testing framework to quantify confidence in the inferred predictions.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Communications Physics
Communications Physics Physics and Astronomy-General Physics and Astronomy
CiteScore
8.40
自引率
3.60%
发文量
276
审稿时长
13 weeks
期刊介绍: Communications Physics is an open access journal from Nature Research publishing high-quality research, reviews and commentary in all areas of the physical sciences. Research papers published by the journal represent significant advances bringing new insight to a specialized area of research in physics. We also aim to provide a community forum for issues of importance to all physicists, regardless of sub-discipline. The scope of the journal covers all areas of experimental, applied, fundamental, and interdisciplinary physical sciences. Primary research published in Communications Physics includes novel experimental results, new techniques or computational methods that may influence the work of others in the sub-discipline. We also consider submissions from adjacent research fields where the central advance of the study is of interest to physicists, for example material sciences, physical chemistry and technologies.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信