使用可解释的连续时间序列特征的变分自编码器

Hendrik Klopries;Andreas Schwung
{"title":"使用可解释的连续时间序列特征的变分自编码器","authors":"Hendrik Klopries;Andreas Schwung","doi":"10.1109/TAI.2025.3545396","DOIUrl":null,"url":null,"abstract":"Machine learning algorithms are driven by data. However, the quantity and quality of data in industries are limited due to multiple process constraints. Generating artificial data and performing a transfer learning task is a common solution to overcome these limitations. Recently, deep generative models have become one of the leading solutions for modeling a given source domain. The main hindrance to using those machine learning approaches is the lack of interpretability. Therefore, we present a novel variational autoencoder approach to generate time series data on a probabilistic latent feature representation and enhance interpretability within the generative model and the output trajectory. We sample selective and parameter values for certain continuous function candidates to assemble the synthetic time series. The sparse design of the generative model enables direct interpretability and matches an estimated posterior distribution of the detected components in the source domain. Through residual stacking, conditionality, and a mixture of prior distributions, we derive a stacked version of the evidence lower bound to learn our network. Tests on synthetic and real industrial datasets underline the performance and interpretability of our generative model. Depending on the model and function candidates, the user can define a trade-off between flexibility and interpretability. Overall, this work presents an innovative interpretable representation of the latent space and further developed evidence lower bound criterion driven by the designed architecture.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 8","pages":"2314-2326"},"PeriodicalIF":0.0000,"publicationDate":"2025-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ITF-VAE: Variational Auto-Encoder Using Interpretable Continuous Time Series Features\",\"authors\":\"Hendrik Klopries;Andreas Schwung\",\"doi\":\"10.1109/TAI.2025.3545396\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Machine learning algorithms are driven by data. However, the quantity and quality of data in industries are limited due to multiple process constraints. Generating artificial data and performing a transfer learning task is a common solution to overcome these limitations. Recently, deep generative models have become one of the leading solutions for modeling a given source domain. The main hindrance to using those machine learning approaches is the lack of interpretability. Therefore, we present a novel variational autoencoder approach to generate time series data on a probabilistic latent feature representation and enhance interpretability within the generative model and the output trajectory. We sample selective and parameter values for certain continuous function candidates to assemble the synthetic time series. The sparse design of the generative model enables direct interpretability and matches an estimated posterior distribution of the detected components in the source domain. Through residual stacking, conditionality, and a mixture of prior distributions, we derive a stacked version of the evidence lower bound to learn our network. Tests on synthetic and real industrial datasets underline the performance and interpretability of our generative model. Depending on the model and function candidates, the user can define a trade-off between flexibility and interpretability. Overall, this work presents an innovative interpretable representation of the latent space and further developed evidence lower bound criterion driven by the designed architecture.\",\"PeriodicalId\":73305,\"journal\":{\"name\":\"IEEE transactions on artificial intelligence\",\"volume\":\"6 8\",\"pages\":\"2314-2326\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-02-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE transactions on artificial intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10902438/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10902438/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

机器学习算法由数据驱动。然而,由于多个过程的约束,工业数据的数量和质量受到限制。生成人工数据和执行迁移学习任务是克服这些限制的常见解决方案。近年来,深度生成模型已成为对给定源域进行建模的主要解决方案之一。使用这些机器学习方法的主要障碍是缺乏可解释性。因此,我们提出了一种新的变分自编码器方法,以概率潜在特征表示生成时间序列数据,并增强生成模型和输出轨迹内的可解释性。我们对某些连续候选函数的选择值和参数值进行采样,以组装合成时间序列。生成模型的稀疏设计使其具有直接的可解释性,并与源域中检测到的成分的估计后验分布相匹配。通过残差叠加、条件性和先验分布的混合,我们推导出证据下界的叠加版本来学习我们的网络。对合成和真实工业数据集的测试强调了我们生成模型的性能和可解释性。根据模型和候选函数,用户可以在灵活性和可解释性之间进行权衡。总体而言,这项工作提出了潜在空间的创新可解释表示,并进一步发展了由设计的建筑驱动的证据下限标准。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
ITF-VAE: Variational Auto-Encoder Using Interpretable Continuous Time Series Features
Machine learning algorithms are driven by data. However, the quantity and quality of data in industries are limited due to multiple process constraints. Generating artificial data and performing a transfer learning task is a common solution to overcome these limitations. Recently, deep generative models have become one of the leading solutions for modeling a given source domain. The main hindrance to using those machine learning approaches is the lack of interpretability. Therefore, we present a novel variational autoencoder approach to generate time series data on a probabilistic latent feature representation and enhance interpretability within the generative model and the output trajectory. We sample selective and parameter values for certain continuous function candidates to assemble the synthetic time series. The sparse design of the generative model enables direct interpretability and matches an estimated posterior distribution of the detected components in the source domain. Through residual stacking, conditionality, and a mixture of prior distributions, we derive a stacked version of the evidence lower bound to learn our network. Tests on synthetic and real industrial datasets underline the performance and interpretability of our generative model. Depending on the model and function candidates, the user can define a trade-off between flexibility and interpretability. Overall, this work presents an innovative interpretable representation of the latent space and further developed evidence lower bound criterion driven by the designed architecture.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
7.70
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信