{"title":"具有不可观测性的物理系统识别的物理可解释浅-深神经网络","authors":"Jingyi Yuan, Yang Weng","doi":"10.1109/ICDM51629.2021.00096","DOIUrl":null,"url":null,"abstract":"The large amount of data collected in complex physical systems allows machine learning models to solve a variety of prediction problems. However, the directly applied learning approaches, especially deep neural networks (DNN), are difficult to balance between universal approximation to minimize error and the interpretability to reveal underlying physical law. Their performance drops even faster with system unobservability (of measurements) issues due to limited measurements. In this paper, we construct the novel physics interpretable shallow-deep neural networks to integrate exact physical interpretation and universal approximation to address the concerns in previous methods. We show that not only the shallow layer of the structural DNN extracts interpretable physical features but also the designed physical-input convex property of the DNN guarantees the true physical function recovery. While input convexity conditions are strict, the proposed model retains the representation capability to universally approximate for the unobservable system regions. We demonstrate its effectiveness by experiments on physical systems. In particular, we implement the proposed model on the forward kinematics and complex power flow reproduction tasks, with or without observability issues. We show that, besides the physical interpretability, our model provides consistently smaller or similar prediction error for system identification, compared to the state-of-art learning methods.","PeriodicalId":320970,"journal":{"name":"2021 IEEE International Conference on Data Mining (ICDM)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Physics Interpretable Shallow-Deep Neural Networks for Physical System Identification with Unobservability\",\"authors\":\"Jingyi Yuan, Yang Weng\",\"doi\":\"10.1109/ICDM51629.2021.00096\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The large amount of data collected in complex physical systems allows machine learning models to solve a variety of prediction problems. However, the directly applied learning approaches, especially deep neural networks (DNN), are difficult to balance between universal approximation to minimize error and the interpretability to reveal underlying physical law. Their performance drops even faster with system unobservability (of measurements) issues due to limited measurements. In this paper, we construct the novel physics interpretable shallow-deep neural networks to integrate exact physical interpretation and universal approximation to address the concerns in previous methods. We show that not only the shallow layer of the structural DNN extracts interpretable physical features but also the designed physical-input convex property of the DNN guarantees the true physical function recovery. While input convexity conditions are strict, the proposed model retains the representation capability to universally approximate for the unobservable system regions. We demonstrate its effectiveness by experiments on physical systems. In particular, we implement the proposed model on the forward kinematics and complex power flow reproduction tasks, with or without observability issues. We show that, besides the physical interpretability, our model provides consistently smaller or similar prediction error for system identification, compared to the state-of-art learning methods.\",\"PeriodicalId\":320970,\"journal\":{\"name\":\"2021 IEEE International Conference on Data Mining (ICDM)\",\"volume\":\"89 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE International Conference on Data Mining (ICDM)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICDM51629.2021.00096\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Data Mining (ICDM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDM51629.2021.00096","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Physics Interpretable Shallow-Deep Neural Networks for Physical System Identification with Unobservability
The large amount of data collected in complex physical systems allows machine learning models to solve a variety of prediction problems. However, the directly applied learning approaches, especially deep neural networks (DNN), are difficult to balance between universal approximation to minimize error and the interpretability to reveal underlying physical law. Their performance drops even faster with system unobservability (of measurements) issues due to limited measurements. In this paper, we construct the novel physics interpretable shallow-deep neural networks to integrate exact physical interpretation and universal approximation to address the concerns in previous methods. We show that not only the shallow layer of the structural DNN extracts interpretable physical features but also the designed physical-input convex property of the DNN guarantees the true physical function recovery. While input convexity conditions are strict, the proposed model retains the representation capability to universally approximate for the unobservable system regions. We demonstrate its effectiveness by experiments on physical systems. In particular, we implement the proposed model on the forward kinematics and complex power flow reproduction tasks, with or without observability issues. We show that, besides the physical interpretability, our model provides consistently smaller or similar prediction error for system identification, compared to the state-of-art learning methods.