Physics Interpretable Shallow-Deep Neural Networks for Physical System Identification with Unobservability

Jingyi Yuan, Yang Weng
{"title":"Physics Interpretable Shallow-Deep Neural Networks for Physical System Identification with Unobservability","authors":"Jingyi Yuan, Yang Weng","doi":"10.1109/ICDM51629.2021.00096","DOIUrl":null,"url":null,"abstract":"The large amount of data collected in complex physical systems allows machine learning models to solve a variety of prediction problems. However, the directly applied learning approaches, especially deep neural networks (DNN), are difficult to balance between universal approximation to minimize error and the interpretability to reveal underlying physical law. Their performance drops even faster with system unobservability (of measurements) issues due to limited measurements. In this paper, we construct the novel physics interpretable shallow-deep neural networks to integrate exact physical interpretation and universal approximation to address the concerns in previous methods. We show that not only the shallow layer of the structural DNN extracts interpretable physical features but also the designed physical-input convex property of the DNN guarantees the true physical function recovery. While input convexity conditions are strict, the proposed model retains the representation capability to universally approximate for the unobservable system regions. We demonstrate its effectiveness by experiments on physical systems. In particular, we implement the proposed model on the forward kinematics and complex power flow reproduction tasks, with or without observability issues. We show that, besides the physical interpretability, our model provides consistently smaller or similar prediction error for system identification, compared to the state-of-art learning methods.","PeriodicalId":320970,"journal":{"name":"2021 IEEE International Conference on Data Mining (ICDM)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Data Mining (ICDM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDM51629.2021.00096","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

The large amount of data collected in complex physical systems allows machine learning models to solve a variety of prediction problems. However, the directly applied learning approaches, especially deep neural networks (DNN), are difficult to balance between universal approximation to minimize error and the interpretability to reveal underlying physical law. Their performance drops even faster with system unobservability (of measurements) issues due to limited measurements. In this paper, we construct the novel physics interpretable shallow-deep neural networks to integrate exact physical interpretation and universal approximation to address the concerns in previous methods. We show that not only the shallow layer of the structural DNN extracts interpretable physical features but also the designed physical-input convex property of the DNN guarantees the true physical function recovery. While input convexity conditions are strict, the proposed model retains the representation capability to universally approximate for the unobservable system regions. We demonstrate its effectiveness by experiments on physical systems. In particular, we implement the proposed model on the forward kinematics and complex power flow reproduction tasks, with or without observability issues. We show that, besides the physical interpretability, our model provides consistently smaller or similar prediction error for system identification, compared to the state-of-art learning methods.
具有不可观测性的物理系统识别的物理可解释浅-深神经网络
在复杂的物理系统中收集的大量数据使机器学习模型能够解决各种预测问题。然而,直接应用的学习方法,特别是深度神经网络(deep neural networks, DNN),很难在最小化误差的通用近似和揭示潜在物理规律的可解释性之间取得平衡。由于有限的测量而导致的系统不可观察性(测量)问题,它们的性能下降得更快。在本文中,我们构建了一种新的物理可解释的浅-深神经网络,将精确物理解释和普遍近似相结合,以解决以往方法中存在的问题。研究表明,不仅结构深度神经网络的浅层提取了可解释的物理特征,而且深度神经网络设计的物理输入凸特性保证了真正的物理功能恢复。在输入凸性条件严格的情况下,该模型保留了对不可观测系统区域进行普遍近似的表示能力。我们通过物理系统的实验证明了它的有效性。特别地,我们在正运动学和复杂的潮流再现任务上实现了所提出的模型,有或没有可观察性问题。我们表明,除了物理可解释性之外,与最先进的学习方法相比,我们的模型为系统识别提供了一致的更小或相似的预测误差。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信