利用负反馈增强曼巴的非线性依赖关系进行时间序列预测

IF 6.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Sijie Xiong , Cheng Tang , Yuanyuan Zhang , Haoling Xiong , Youhao Xu , Atsushi Shimada
{"title":"利用负反馈增强曼巴的非线性依赖关系进行时间序列预测","authors":"Sijie Xiong ,&nbsp;Cheng Tang ,&nbsp;Yuanyuan Zhang ,&nbsp;Haoling Xiong ,&nbsp;Youhao Xu ,&nbsp;Atsushi Shimada","doi":"10.1016/j.asoc.2025.113758","DOIUrl":null,"url":null,"abstract":"<div><div>Mamba is a rising model designed to distill complex patterns from historical data, providing predictive capabilities for time series forecasting tasks. Mamba’s similarity to linear-based models has been criticized due to its limited ability to capture nonlinear dependencies. In this work, we propose a novel model named Embedding <strong><em>C</em></strong>hannel Attention <strong><em>M</em></strong>aclaurin <strong><em>E</em></strong>instein <strong><em>Mamba</em></strong> (CME-Mamba<span><span><sup>1</sup></span></span>.) based on Mamba framework, with both Embedding Channel Attention and Maclaurin mechanisms incorporated. To further address gradient vanishing issues, we integrate Einstein FFT algorithms, ensuring robust performance against abnormal behaviors of Mamba-based architectures. Extensive experiments conducted on 11 real-world datasets with different numbers of variates, domain focus and granularity, reveal that CME-Mamba achieves state-of-the-art performance in both MSE and MAE, while maintaining reasonable memory efficiency and low time cost. The robustness and credibility of all results are substantiated by a comprehensive convergence and stability analysis. Statistically, consolidated by the Friedman Nonparametric Test and the Wilcoxon Signed-Rank Test, CME-Mamba ranks the first place with significance over counterparts. In addition, in terms of time and memory analysis, CME-Mamba is among the top three models for time and memory efficiency. Despite this, our results further demonstrate that the main contributor is the Embedding Channel Attention Block, which greatly enhances nonlinear dependencies over datasets. The Einstein FFT Block effectively suppresses gradient vanishing occurrences and contributes considerably to performance improvements, driving CME-Mamba both stable and promising. Moreover, the Maclaurin Block based on negative feedback is asymptotically stable without additional gradient vanishing issues and pioneered in achieving synergies with other blocks and greatly enhances nonlinear dependencies. With enhanced nonlinear dependencies generated from the synergy effect of all the three blocks, CME-Mamba grows excellent to uncover complex paradigms and predict future states in various domains, especially improving the performance for periodic and high-variate situations, such as traffic flow management (<span><math><mrow><mo>≈</mo><mo>+</mo><mn>8</mn><mtext>%</mtext></mrow></math></span>), electricity predictions(<span><math><mrow><mo>≈</mo><mo>+</mo><mn>6</mn><mtext>%</mtext></mrow></math></span>).</div></div>","PeriodicalId":50737,"journal":{"name":"Applied Soft Computing","volume":"184 ","pages":"Article 113758"},"PeriodicalIF":6.6000,"publicationDate":"2025-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enhancing nonlinear dependencies of Mamba via negative feedback for time series forecasting\",\"authors\":\"Sijie Xiong ,&nbsp;Cheng Tang ,&nbsp;Yuanyuan Zhang ,&nbsp;Haoling Xiong ,&nbsp;Youhao Xu ,&nbsp;Atsushi Shimada\",\"doi\":\"10.1016/j.asoc.2025.113758\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Mamba is a rising model designed to distill complex patterns from historical data, providing predictive capabilities for time series forecasting tasks. Mamba’s similarity to linear-based models has been criticized due to its limited ability to capture nonlinear dependencies. In this work, we propose a novel model named Embedding <strong><em>C</em></strong>hannel Attention <strong><em>M</em></strong>aclaurin <strong><em>E</em></strong>instein <strong><em>Mamba</em></strong> (CME-Mamba<span><span><sup>1</sup></span></span>.) based on Mamba framework, with both Embedding Channel Attention and Maclaurin mechanisms incorporated. To further address gradient vanishing issues, we integrate Einstein FFT algorithms, ensuring robust performance against abnormal behaviors of Mamba-based architectures. Extensive experiments conducted on 11 real-world datasets with different numbers of variates, domain focus and granularity, reveal that CME-Mamba achieves state-of-the-art performance in both MSE and MAE, while maintaining reasonable memory efficiency and low time cost. The robustness and credibility of all results are substantiated by a comprehensive convergence and stability analysis. Statistically, consolidated by the Friedman Nonparametric Test and the Wilcoxon Signed-Rank Test, CME-Mamba ranks the first place with significance over counterparts. In addition, in terms of time and memory analysis, CME-Mamba is among the top three models for time and memory efficiency. Despite this, our results further demonstrate that the main contributor is the Embedding Channel Attention Block, which greatly enhances nonlinear dependencies over datasets. The Einstein FFT Block effectively suppresses gradient vanishing occurrences and contributes considerably to performance improvements, driving CME-Mamba both stable and promising. Moreover, the Maclaurin Block based on negative feedback is asymptotically stable without additional gradient vanishing issues and pioneered in achieving synergies with other blocks and greatly enhances nonlinear dependencies. With enhanced nonlinear dependencies generated from the synergy effect of all the three blocks, CME-Mamba grows excellent to uncover complex paradigms and predict future states in various domains, especially improving the performance for periodic and high-variate situations, such as traffic flow management (<span><math><mrow><mo>≈</mo><mo>+</mo><mn>8</mn><mtext>%</mtext></mrow></math></span>), electricity predictions(<span><math><mrow><mo>≈</mo><mo>+</mo><mn>6</mn><mtext>%</mtext></mrow></math></span>).</div></div>\",\"PeriodicalId\":50737,\"journal\":{\"name\":\"Applied Soft Computing\",\"volume\":\"184 \",\"pages\":\"Article 113758\"},\"PeriodicalIF\":6.6000,\"publicationDate\":\"2025-08-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Soft Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1568494625010713\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Soft Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1568494625010713","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

Mamba是一个新兴的模型,旨在从历史数据中提取复杂的模式,为时间序列预测任务提供预测能力。Mamba与基于线性的模型的相似性由于其捕获非线性依赖关系的能力有限而受到批评。在本研究中,我们提出了一个基于曼巴框架的嵌入通道注意麦克劳林爱因斯坦曼巴(CME-Mamba1.)模型,该模型结合了嵌入通道注意和麦克劳林机制。为了进一步解决梯度消失问题,我们集成了爱因斯坦FFT算法,确保对基于mamba架构的异常行为的鲁棒性能。在11个具有不同变量数量、领域焦点和粒度的真实数据集上进行的大量实验表明,CME-Mamba在MSE和MAE方面都取得了最先进的性能,同时保持了合理的内存效率和较低的时间成本。通过全面的收敛性和稳定性分析,证明了所有结果的鲁棒性和可信性。统计上,经Friedman非参数检验和Wilcoxon Signed-Rank检验巩固,CME-Mamba的显著性排名第一。此外,在时间和内存分析方面,CME-Mamba在时间和内存效率方面排名前三。尽管如此,我们的结果进一步表明,主要贡献者是嵌入通道注意块,它大大增强了数据集上的非线性依赖关系。爱因斯坦FFT块有效地抑制了梯度消失的发生,并大大有助于性能改进,使CME-Mamba既稳定又有前途。此外,基于负反馈的Maclaurin块是渐近稳定的,没有额外的梯度消失问题,并且率先实现了与其他块的协同作用,大大增强了非线性依赖性。由于三个区块的协同效应增强了非线性依赖关系,CME-Mamba在揭示复杂范式和预测各个领域的未来状态方面表现出色,特别是在周期性和高变量情况下的性能提高,例如交通流量管理(≈+8%),电力预测(≈+6%)。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Enhancing nonlinear dependencies of Mamba via negative feedback for time series forecasting
Mamba is a rising model designed to distill complex patterns from historical data, providing predictive capabilities for time series forecasting tasks. Mamba’s similarity to linear-based models has been criticized due to its limited ability to capture nonlinear dependencies. In this work, we propose a novel model named Embedding Channel Attention Maclaurin Einstein Mamba (CME-Mamba1.) based on Mamba framework, with both Embedding Channel Attention and Maclaurin mechanisms incorporated. To further address gradient vanishing issues, we integrate Einstein FFT algorithms, ensuring robust performance against abnormal behaviors of Mamba-based architectures. Extensive experiments conducted on 11 real-world datasets with different numbers of variates, domain focus and granularity, reveal that CME-Mamba achieves state-of-the-art performance in both MSE and MAE, while maintaining reasonable memory efficiency and low time cost. The robustness and credibility of all results are substantiated by a comprehensive convergence and stability analysis. Statistically, consolidated by the Friedman Nonparametric Test and the Wilcoxon Signed-Rank Test, CME-Mamba ranks the first place with significance over counterparts. In addition, in terms of time and memory analysis, CME-Mamba is among the top three models for time and memory efficiency. Despite this, our results further demonstrate that the main contributor is the Embedding Channel Attention Block, which greatly enhances nonlinear dependencies over datasets. The Einstein FFT Block effectively suppresses gradient vanishing occurrences and contributes considerably to performance improvements, driving CME-Mamba both stable and promising. Moreover, the Maclaurin Block based on negative feedback is asymptotically stable without additional gradient vanishing issues and pioneered in achieving synergies with other blocks and greatly enhances nonlinear dependencies. With enhanced nonlinear dependencies generated from the synergy effect of all the three blocks, CME-Mamba grows excellent to uncover complex paradigms and predict future states in various domains, especially improving the performance for periodic and high-variate situations, such as traffic flow management (+8%), electricity predictions(+6%).
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Applied Soft Computing
Applied Soft Computing 工程技术-计算机:跨学科应用
CiteScore
15.80
自引率
6.90%
发文量
874
审稿时长
10.9 months
期刊介绍: Applied Soft Computing is an international journal promoting an integrated view of soft computing to solve real life problems.The focus is to publish the highest quality research in application and convergence of the areas of Fuzzy Logic, Neural Networks, Evolutionary Computing, Rough Sets and other similar techniques to address real world complexities. Applied Soft Computing is a rolling publication: articles are published as soon as the editor-in-chief has accepted them. Therefore, the web site will continuously be updated with new articles and the publication time will be short.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信