A dual-reconstruction self-rectification framework with momentum memory-augmented network for multivariate time series anomaly detection

IF 6.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Bing Xue , Xin Gao , Heping Lu , Baofeng Li , Feng Zhai , Meng Xu , Taizhi Wang , Jiawen Lu
{"title":"A dual-reconstruction self-rectification framework with momentum memory-augmented network for multivariate time series anomaly detection","authors":"Bing Xue ,&nbsp;Xin Gao ,&nbsp;Heping Lu ,&nbsp;Baofeng Li ,&nbsp;Feng Zhai ,&nbsp;Meng Xu ,&nbsp;Taizhi Wang ,&nbsp;Jiawen Lu","doi":"10.1016/j.asoc.2025.113558","DOIUrl":null,"url":null,"abstract":"<div><div>The discrepancy between the actual contaminated data and the normality assumption poses a serious challenge to existing methods that rely on clean training data. For methods that consider contamination, the mainstream model-level methods with memory-augmented structures struggle with biased similarity measures and fail to utilize historical information, leading to inaccurate reconstruction of latent variables. Most training-level methods may confuse contaminated data with hard-to-learn normal data, affecting the model’s ability to learn normal patterns. Moreover, there is a lack of using adjustment loss to effectively constrain model-level methods to learn normal data while suppressing contaminated data, which limits the further improvement of anomaly detection performance. This paper proposes a <strong>D</strong>ual-Reconstruction Self-<strong>R</strong>ectification framework with <strong>Mo</strong>mentum <strong>Me</strong>mory-augmented network based on Transformer (<strong>DRMoMe</strong>) for multivariate time series anomaly detection. At the model level, a momentum memory module based on Transformer is proposed, which employs the momentum-updated framework to align the representation space and designs the multihead-attention mechanism with the similarity-based update strategy to ensure the accuracy and diversity of the memory vectors. At the training level, this paper designs a self-rectification framework, which uses the difference between dual-reconstruction paths as the loss adjustment weights to adjust the model’s learning dynamically. Additionally, the method uses the characteristic of the memory module to amplify the weight difference between the contaminated and normal data, effectively integrating the model-level and training-level approach to help the model focus on learning the normal pattern. The DRMoMe outperforms 21 state-of-the-art baselines in experiments conducted on five benchmark datasets from different domains.</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":"182 ","pages":"Article 113558"},"PeriodicalIF":6.6000,"publicationDate":"2025-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1568494625008695","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

The discrepancy between the actual contaminated data and the normality assumption poses a serious challenge to existing methods that rely on clean training data. For methods that consider contamination, the mainstream model-level methods with memory-augmented structures struggle with biased similarity measures and fail to utilize historical information, leading to inaccurate reconstruction of latent variables. Most training-level methods may confuse contaminated data with hard-to-learn normal data, affecting the model’s ability to learn normal patterns. Moreover, there is a lack of using adjustment loss to effectively constrain model-level methods to learn normal data while suppressing contaminated data, which limits the further improvement of anomaly detection performance. This paper proposes a Dual-Reconstruction Self-Rectification framework with Momentum Memory-augmented network based on Transformer (DRMoMe) for multivariate time series anomaly detection. At the model level, a momentum memory module based on Transformer is proposed, which employs the momentum-updated framework to align the representation space and designs the multihead-attention mechanism with the similarity-based update strategy to ensure the accuracy and diversity of the memory vectors. At the training level, this paper designs a self-rectification framework, which uses the difference between dual-reconstruction paths as the loss adjustment weights to adjust the model’s learning dynamically. Additionally, the method uses the characteristic of the memory module to amplify the weight difference between the contaminated and normal data, effectively integrating the model-level and training-level approach to help the model focus on learning the normal pattern. The DRMoMe outperforms 21 state-of-the-art baselines in experiments conducted on five benchmark datasets from different domains.
基于动量记忆增强网络的双重构自校正框架多变量时间序列异常检测
实际污染数据与正态性假设之间的差异对依赖干净训练数据的现有方法提出了严峻的挑战。对于考虑污染的方法,具有记忆增强结构的主流模型级方法与有偏差的相似性度量相斗争,并且不能利用历史信息,导致潜在变量的不准确重建。大多数训练级方法可能会将污染数据与难以学习的正常数据混淆,从而影响模型学习正常模式的能力。此外,缺乏利用调整损失来有效约束模型级方法在抑制污染数据的同时学习正常数据,限制了异常检测性能的进一步提高。针对多元时间序列异常检测问题,提出了一种基于变压器的动量记忆增强网络双重构自纠偏框架。在模型层面,提出了一种基于Transformer的动量记忆模块,该模块采用动量更新框架对齐表示空间,并采用基于相似度的更新策略设计多头关注机制,保证了记忆向量的准确性和多样性。在训练层面,设计了自校正框架,利用双重构路径的差值作为损失调整权值,动态调整模型的学习。此外,该方法利用记忆模块的特性放大污染数据与正常数据之间的权重差,有效地将模型级和训练级方法相结合,帮助模型专注于正常模式的学习。在来自不同领域的五个基准数据集上进行的实验中,DRMoMe优于21个最先进的基线。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Applied Soft Computing
Applied Soft Computing 工程技术-计算机:跨学科应用
CiteScore
15.80
自引率
6.90%
发文量
874
审稿时长
10.9 months
期刊介绍: Applied Soft Computing is an international journal promoting an integrated view of soft computing to solve real life problems.The focus is to publish the highest quality research in application and convergence of the areas of Fuzzy Logic, Neural Networks, Evolutionary Computing, Rough Sets and other similar techniques to address real world complexities. Applied Soft Computing is a rolling publication: articles are published as soon as the editor-in-chief has accepted them. Therefore, the web site will continuously be updated with new articles and the publication time will be short.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信