{"title":"Multiscale spatial-temporal transformer with consistency representation learning for multivariate time series classification","authors":"Wei Wu, Feiyue Qiu, Liping Wang, Yanxiu Liu","doi":"10.1002/cpe.8234","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>Multivariate time series classification holds significant importance in fields such as healthcare, energy management, and industrial manufacturing. Existing research focuses on capturing temporal changes or calculating time similarities to accomplish classification tasks. However, as the state of the system changes, capturing spatial-temporal consistency within multivariate time series is key to the ability of the model to classify accurately. This paper proposes the MSTformer model, specifically designed for multivariate time series classification tasks. Based on the Transformer architecture, this model uniquely focuses on multiscale information across both time and feature dimensions. The encoder, through a designed learnable multiscale attention mechanism, divides data into sequences of varying temporal scales to learn multiscale temporal features. The decoder, which receives the spatial view of the data, utilizes a dynamic scale attention mechanism to learn spatial-temporal consistency in a one-dimensional space. In addition, this paper proposes an adaptive aggregation mechanism to synchronize and combine the outputs of the encoder and decoder. It also introduces a multiscale 2D separable convolution designed to learn spatial-temporal consistency in two-dimensional space, enhancing the ability of the model to learn spatial-temporal consistency representation. Extensive experiments were conducted on 30 datasets, where the MSTformer outperformed other models with an average accuracy rate of 85.6%. Ablation studies further demonstrate the reliability and stability of MSTformer.</p>\n </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"36 27","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2024-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Concurrency and Computation-Practice & Experience","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/cpe.8234","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 0
Abstract
Multivariate time series classification holds significant importance in fields such as healthcare, energy management, and industrial manufacturing. Existing research focuses on capturing temporal changes or calculating time similarities to accomplish classification tasks. However, as the state of the system changes, capturing spatial-temporal consistency within multivariate time series is key to the ability of the model to classify accurately. This paper proposes the MSTformer model, specifically designed for multivariate time series classification tasks. Based on the Transformer architecture, this model uniquely focuses on multiscale information across both time and feature dimensions. The encoder, through a designed learnable multiscale attention mechanism, divides data into sequences of varying temporal scales to learn multiscale temporal features. The decoder, which receives the spatial view of the data, utilizes a dynamic scale attention mechanism to learn spatial-temporal consistency in a one-dimensional space. In addition, this paper proposes an adaptive aggregation mechanism to synchronize and combine the outputs of the encoder and decoder. It also introduces a multiscale 2D separable convolution designed to learn spatial-temporal consistency in two-dimensional space, enhancing the ability of the model to learn spatial-temporal consistency representation. Extensive experiments were conducted on 30 datasets, where the MSTformer outperformed other models with an average accuracy rate of 85.6%. Ablation studies further demonstrate the reliability and stability of MSTformer.
期刊介绍:
Concurrency and Computation: Practice and Experience (CCPE) publishes high-quality, original research papers, and authoritative research review papers, in the overlapping fields of:
Parallel and distributed computing;
High-performance computing;
Computational and data science;
Artificial intelligence and machine learning;
Big data applications, algorithms, and systems;
Network science;
Ontologies and semantics;
Security and privacy;
Cloud/edge/fog computing;
Green computing; and
Quantum computing.