Bing Xue , Xin Gao , Heping Lu , Baofeng Li , Feng Zhai , Meng Xu , Taizhi Wang , Jiawen Lu
{"title":"A dual-reconstruction self-rectification framework with momentum memory-augmented network for multivariate time series anomaly detection","authors":"Bing Xue , Xin Gao , Heping Lu , Baofeng Li , Feng Zhai , Meng Xu , Taizhi Wang , Jiawen Lu","doi":"10.1016/j.asoc.2025.113558","DOIUrl":null,"url":null,"abstract":"<div><div>The discrepancy between the actual contaminated data and the normality assumption poses a serious challenge to existing methods that rely on clean training data. For methods that consider contamination, the mainstream model-level methods with memory-augmented structures struggle with biased similarity measures and fail to utilize historical information, leading to inaccurate reconstruction of latent variables. Most training-level methods may confuse contaminated data with hard-to-learn normal data, affecting the model’s ability to learn normal patterns. Moreover, there is a lack of using adjustment loss to effectively constrain model-level methods to learn normal data while suppressing contaminated data, which limits the further improvement of anomaly detection performance. This paper proposes a <strong>D</strong>ual-Reconstruction Self-<strong>R</strong>ectification framework with <strong>Mo</strong>mentum <strong>Me</strong>mory-augmented network based on Transformer (<strong>DRMoMe</strong>) for multivariate time series anomaly detection. At the model level, a momentum memory module based on Transformer is proposed, which employs the momentum-updated framework to align the representation space and designs the multihead-attention mechanism with the similarity-based update strategy to ensure the accuracy and diversity of the memory vectors. At the training level, this paper designs a self-rectification framework, which uses the difference between dual-reconstruction paths as the loss adjustment weights to adjust the model’s learning dynamically. Additionally, the method uses the characteristic of the memory module to amplify the weight difference between the contaminated and normal data, effectively integrating the model-level and training-level approach to help the model focus on learning the normal pattern. The DRMoMe outperforms 21 state-of-the-art baselines in experiments conducted on five benchmark datasets from different domains.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":"182 ","pages":"Article 113558"},"PeriodicalIF":6.6000,"publicationDate":"2025-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1568494625008695","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The discrepancy between the actual contaminated data and the normality assumption poses a serious challenge to existing methods that rely on clean training data. For methods that consider contamination, the mainstream model-level methods with memory-augmented structures struggle with biased similarity measures and fail to utilize historical information, leading to inaccurate reconstruction of latent variables. Most training-level methods may confuse contaminated data with hard-to-learn normal data, affecting the model’s ability to learn normal patterns. Moreover, there is a lack of using adjustment loss to effectively constrain model-level methods to learn normal data while suppressing contaminated data, which limits the further improvement of anomaly detection performance. This paper proposes a Dual-Reconstruction Self-Rectification framework with Momentum Memory-augmented network based on Transformer (DRMoMe) for multivariate time series anomaly detection. At the model level, a momentum memory module based on Transformer is proposed, which employs the momentum-updated framework to align the representation space and designs the multihead-attention mechanism with the similarity-based update strategy to ensure the accuracy and diversity of the memory vectors. At the training level, this paper designs a self-rectification framework, which uses the difference between dual-reconstruction paths as the loss adjustment weights to adjust the model’s learning dynamically. Additionally, the method uses the characteristic of the memory module to amplify the weight difference between the contaminated and normal data, effectively integrating the model-level and training-level approach to help the model focus on learning the normal pattern. The DRMoMe outperforms 21 state-of-the-art baselines in experiments conducted on five benchmark datasets from different domains.
期刊介绍:
Applied Soft Computing is an international journal promoting an integrated view of soft computing to solve real life problems.The focus is to publish the highest quality research in application and convergence of the areas of Fuzzy Logic, Neural Networks, Evolutionary Computing, Rough Sets and other similar techniques to address real world complexities.
Applied Soft Computing is a rolling publication: articles are published as soon as the editor-in-chief has accepted them. Therefore, the web site will continuously be updated with new articles and the publication time will be short.