Multiscale spatial-temporal transformer with consistency representation learning for multivariate time series classification

IF 1.5 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING
Wei Wu, Feiyue Qiu, Liping Wang, Yanxiu Liu
{"title":"Multiscale spatial-temporal transformer with consistency representation learning for multivariate time series classification","authors":"Wei Wu,&nbsp;Feiyue Qiu,&nbsp;Liping Wang,&nbsp;Yanxiu Liu","doi":"10.1002/cpe.8234","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>Multivariate time series classification holds significant importance in fields such as healthcare, energy management, and industrial manufacturing. Existing research focuses on capturing temporal changes or calculating time similarities to accomplish classification tasks. However, as the state of the system changes, capturing spatial-temporal consistency within multivariate time series is key to the ability of the model to classify accurately. This paper proposes the MSTformer model, specifically designed for multivariate time series classification tasks. Based on the Transformer architecture, this model uniquely focuses on multiscale information across both time and feature dimensions. The encoder, through a designed learnable multiscale attention mechanism, divides data into sequences of varying temporal scales to learn multiscale temporal features. The decoder, which receives the spatial view of the data, utilizes a dynamic scale attention mechanism to learn spatial-temporal consistency in a one-dimensional space. In addition, this paper proposes an adaptive aggregation mechanism to synchronize and combine the outputs of the encoder and decoder. It also introduces a multiscale 2D separable convolution designed to learn spatial-temporal consistency in two-dimensional space, enhancing the ability of the model to learn spatial-temporal consistency representation. Extensive experiments were conducted on 30 datasets, where the MSTformer outperformed other models with an average accuracy rate of 85.6%. Ablation studies further demonstrate the reliability and stability of MSTformer.</p>\n </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"36 27","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2024-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Concurrency and Computation-Practice & Experience","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/cpe.8234","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 0

Abstract

Multivariate time series classification holds significant importance in fields such as healthcare, energy management, and industrial manufacturing. Existing research focuses on capturing temporal changes or calculating time similarities to accomplish classification tasks. However, as the state of the system changes, capturing spatial-temporal consistency within multivariate time series is key to the ability of the model to classify accurately. This paper proposes the MSTformer model, specifically designed for multivariate time series classification tasks. Based on the Transformer architecture, this model uniquely focuses on multiscale information across both time and feature dimensions. The encoder, through a designed learnable multiscale attention mechanism, divides data into sequences of varying temporal scales to learn multiscale temporal features. The decoder, which receives the spatial view of the data, utilizes a dynamic scale attention mechanism to learn spatial-temporal consistency in a one-dimensional space. In addition, this paper proposes an adaptive aggregation mechanism to synchronize and combine the outputs of the encoder and decoder. It also introduces a multiscale 2D separable convolution designed to learn spatial-temporal consistency in two-dimensional space, enhancing the ability of the model to learn spatial-temporal consistency representation. Extensive experiments were conducted on 30 datasets, where the MSTformer outperformed other models with an average accuracy rate of 85.6%. Ablation studies further demonstrate the reliability and stability of MSTformer.

多尺度时空变换器与多变量时间序列分类的一致性表示学习
多变量时间序列分类在医疗保健、能源管理和工业制造等领域具有重要意义。现有研究侧重于捕捉时间变化或计算时间相似性来完成分类任务。然而,随着系统状态的变化,捕捉多变量时间序列中的时空一致性是模型能否准确分类的关键。本文提出了专为多变量时间序列分类任务设计的 MSTformer 模型。基于 Transformer 架构,该模型独特地关注时间维度和特征维度的多尺度信息。编码器通过设计的可学习多尺度关注机制,将数据划分为不同时间尺度的序列,以学习多尺度时间特征。解码器接收数据的空间视图,利用动态尺度注意机制学习一维空间中的时空一致性。此外,本文还提出了一种自适应聚合机制,用于同步和合并编码器和解码器的输出。本文还引入了一种多尺度二维可分离卷积,旨在学习二维空间中的时空一致性,从而增强模型学习时空一致性表征的能力。在 30 个数据集上进行了广泛的实验,MSTformer 的表现优于其他模型,平均准确率达到 85.6%。消融研究进一步证明了 MSTformer 的可靠性和稳定性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Concurrency and Computation-Practice & Experience
Concurrency and Computation-Practice & Experience 工程技术-计算机:理论方法
CiteScore
5.00
自引率
10.00%
发文量
664
审稿时长
9.6 months
期刊介绍: Concurrency and Computation: Practice and Experience (CCPE) publishes high-quality, original research papers, and authoritative research review papers, in the overlapping fields of: Parallel and distributed computing; High-performance computing; Computational and data science; Artificial intelligence and machine learning; Big data applications, algorithms, and systems; Network science; Ontologies and semantics; Security and privacy; Cloud/edge/fog computing; Green computing; and Quantum computing.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信