基于深度学习 LSTM 的 10.7 厘米太阳射电通量预报 45 天方法

IF 1.9 4区 物理与天体物理 Q2 ASTRONOMY & ASTROPHYSICS
G. Jerse , A. Marcucci
{"title":"基于深度学习 LSTM 的 10.7 厘米太阳射电通量预报 45 天方法","authors":"G. Jerse ,&nbsp;A. Marcucci","doi":"10.1016/j.ascom.2024.100786","DOIUrl":null,"url":null,"abstract":"<div><p>Accurate forecasting of F10.7 index is important on short, medium and long-term timescales since F10.7 is an excellent proxy of solar activity and it plays an important role within the Space Weather framework. The analysis of the signatures of transient solar radio emission and its prediction are a challenging task as the underpinning physical processes are typically nonlinear, non-stationary and chaotic. In this paper we want to present three Deep Learning approaches for the daily forecasting of the adjusted F10.7 solar radio flux up to 45 days, using a family of Long Short Term Memory (LSTM) based models. We investigated two novel hybrid architectures: the LSTM model used in combination with Fast Iterative Filtering as decomposition algorithm (FIF-LSTM) and a method based on Multi-Head-Attention architecture (FIF-LSTM-MHA). FIF is a robust decomposition signal method very suitable for analyzing non-linear and non-stationary time series and it is used to separate the original time series into different oscillation components according to frequency, derived without leaving the time domain before to be fed into the neural network. The Attention mechanism is able to keep track of long-term dependencies in data sequences and improve the computational efficiency of the prediction model by reducing the effect of irrelevant information, mimicking human attention and selecting the most critical input. Our comparative analysis evaluated the models’ performance for different time lags and solar activity levels. The results indicated that the hybrid models achieve better performance than the LSTM model for mid-range F10.7 predictions while the LSTM achieves better performance within the first few time lags. FIF-LSTM-MHA gives more promising output for longer forecasts since it tends to smooth the prediction curve due to the peculiarity of the Attention module to discard less relevant features of the time series and highlight the global trend.</p></div>","PeriodicalId":48757,"journal":{"name":"Astronomy and Computing","volume":"46 ","pages":"Article 100786"},"PeriodicalIF":1.9000,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2213133724000015/pdfft?md5=702c624fbdced03b7eb08d858ec2dc15&pid=1-s2.0-S2213133724000015-main.pdf","citationCount":"0","resultStr":"{\"title\":\"Deep Learning LSTM-based approaches for 10.7 cm solar radio flux forecasting up to 45-days\",\"authors\":\"G. Jerse ,&nbsp;A. Marcucci\",\"doi\":\"10.1016/j.ascom.2024.100786\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Accurate forecasting of F10.7 index is important on short, medium and long-term timescales since F10.7 is an excellent proxy of solar activity and it plays an important role within the Space Weather framework. The analysis of the signatures of transient solar radio emission and its prediction are a challenging task as the underpinning physical processes are typically nonlinear, non-stationary and chaotic. In this paper we want to present three Deep Learning approaches for the daily forecasting of the adjusted F10.7 solar radio flux up to 45 days, using a family of Long Short Term Memory (LSTM) based models. We investigated two novel hybrid architectures: the LSTM model used in combination with Fast Iterative Filtering as decomposition algorithm (FIF-LSTM) and a method based on Multi-Head-Attention architecture (FIF-LSTM-MHA). FIF is a robust decomposition signal method very suitable for analyzing non-linear and non-stationary time series and it is used to separate the original time series into different oscillation components according to frequency, derived without leaving the time domain before to be fed into the neural network. The Attention mechanism is able to keep track of long-term dependencies in data sequences and improve the computational efficiency of the prediction model by reducing the effect of irrelevant information, mimicking human attention and selecting the most critical input. Our comparative analysis evaluated the models’ performance for different time lags and solar activity levels. The results indicated that the hybrid models achieve better performance than the LSTM model for mid-range F10.7 predictions while the LSTM achieves better performance within the first few time lags. FIF-LSTM-MHA gives more promising output for longer forecasts since it tends to smooth the prediction curve due to the peculiarity of the Attention module to discard less relevant features of the time series and highlight the global trend.</p></div>\",\"PeriodicalId\":48757,\"journal\":{\"name\":\"Astronomy and Computing\",\"volume\":\"46 \",\"pages\":\"Article 100786\"},\"PeriodicalIF\":1.9000,\"publicationDate\":\"2024-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.sciencedirect.com/science/article/pii/S2213133724000015/pdfft?md5=702c624fbdced03b7eb08d858ec2dc15&pid=1-s2.0-S2213133724000015-main.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Astronomy and Computing\",\"FirstCategoryId\":\"101\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2213133724000015\",\"RegionNum\":4,\"RegionCategory\":\"物理与天体物理\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ASTRONOMY & ASTROPHYSICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Astronomy and Computing","FirstCategoryId":"101","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2213133724000015","RegionNum":4,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ASTRONOMY & ASTROPHYSICS","Score":null,"Total":0}
引用次数: 0

摘要

准确预报 F10.7 指数在短期、中期和长期时间尺度上都很重要,因为 F10.7 是太阳活动的极佳代表,在空间天气框架内发挥着重要作用。对瞬态太阳射电辐射特征的分析及其预测是一项具有挑战性的任务,因为其基础物理过程通常是非线性、非稳态和混沌的。在本文中,我们将介绍三种深度学习方法,利用基于长短期记忆(LSTM)的模型系列,对调整后的 F10.7 太阳射电通量进行长达 45 天的每日预测。我们研究了两种新型混合架构:结合快速迭代滤波分解算法(FIF-LSTM)使用的 LSTM 模型和基于多头注意力架构(FIF-LSTM-MHA)的方法。FIF 是一种稳健的分解信号方法,非常适合分析非线性和非稳态时间序列,它用于将原始时间序列按照频率分离成不同的振荡成分,这些成分在输入神经网络之前不会离开时域。注意力机制能够跟踪数据序列中的长期依赖关系,并通过减少无关信息的影响、模仿人类注意力和选择最关键的输入来提高预测模型的计算效率。我们的对比分析评估了模型在不同时滞和太阳活动水平下的性能。结果表明,混合模型在中档 F10.7 预测方面的性能优于 LSTM 模型,而 LSTM 在前几个时滞内的性能更好。FIF-LSTM-MHA 在较长时间的预测中提供了更有前途的输出,因为它倾向于平滑预测曲线,这是由于注意模块的特殊性决定的,它舍弃了时间序列中不那么相关的特征,而突出了全球趋势。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Deep Learning LSTM-based approaches for 10.7 cm solar radio flux forecasting up to 45-days

Accurate forecasting of F10.7 index is important on short, medium and long-term timescales since F10.7 is an excellent proxy of solar activity and it plays an important role within the Space Weather framework. The analysis of the signatures of transient solar radio emission and its prediction are a challenging task as the underpinning physical processes are typically nonlinear, non-stationary and chaotic. In this paper we want to present three Deep Learning approaches for the daily forecasting of the adjusted F10.7 solar radio flux up to 45 days, using a family of Long Short Term Memory (LSTM) based models. We investigated two novel hybrid architectures: the LSTM model used in combination with Fast Iterative Filtering as decomposition algorithm (FIF-LSTM) and a method based on Multi-Head-Attention architecture (FIF-LSTM-MHA). FIF is a robust decomposition signal method very suitable for analyzing non-linear and non-stationary time series and it is used to separate the original time series into different oscillation components according to frequency, derived without leaving the time domain before to be fed into the neural network. The Attention mechanism is able to keep track of long-term dependencies in data sequences and improve the computational efficiency of the prediction model by reducing the effect of irrelevant information, mimicking human attention and selecting the most critical input. Our comparative analysis evaluated the models’ performance for different time lags and solar activity levels. The results indicated that the hybrid models achieve better performance than the LSTM model for mid-range F10.7 predictions while the LSTM achieves better performance within the first few time lags. FIF-LSTM-MHA gives more promising output for longer forecasts since it tends to smooth the prediction curve due to the peculiarity of the Attention module to discard less relevant features of the time series and highlight the global trend.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Astronomy and Computing
Astronomy and Computing ASTRONOMY & ASTROPHYSICSCOMPUTER SCIENCE,-COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS
CiteScore
4.10
自引率
8.00%
发文量
67
期刊介绍: Astronomy and Computing is a peer-reviewed journal that focuses on the broad area between astronomy, computer science and information technology. The journal aims to publish the work of scientists and (software) engineers in all aspects of astronomical computing, including the collection, analysis, reduction, visualisation, preservation and dissemination of data, and the development of astronomical software and simulations. The journal covers applications for academic computer science techniques to astronomy, as well as novel applications of information technologies within astronomy.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信